ansible-playbook 2.9.27
  config file = None
  configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
  ansible python module location = /usr/local/lib/python3.9/site-packages/ansible
  executable location = /usr/local/bin/ansible-playbook
  python version = 3.9.19 (main, May 16 2024, 11:40:09) [GCC 8.5.0 20210514 (Red Hat 8.5.0-22)]
No config file found; using defaults
[WARNING]: running playbook inside collection fedora.linux_system_roles
Skipping callback 'actionable', as we already have a stdout callback.
Skipping callback 'counter_enabled', as we already have a stdout callback.
Skipping callback 'debug', as we already have a stdout callback.
Skipping callback 'dense', as we already have a stdout callback.
Skipping callback 'dense', as we already have a stdout callback.
Skipping callback 'full_skip', as we already have a stdout callback.
Skipping callback 'json', as we already have a stdout callback.
Skipping callback 'minimal', as we already have a stdout callback.
Skipping callback 'null', as we already have a stdout callback.
Skipping callback 'oneline', as we already have a stdout callback.
Skipping callback 'selective', as we already have a stdout callback.
Skipping callback 'skippy', as we already have a stdout callback.
Skipping callback 'stderr', as we already have a stdout callback.
Skipping callback 'unixy', as we already have a stdout callback.
Skipping callback 'yaml', as we already have a stdout callback.

PLAYBOOK: tests_resize.yml *****************************************************
1 plays in /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml

PLAY [Test resize] *************************************************************

TASK [Gathering Facts] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:2
Wednesday 11 December 2024  10:07:25 -0500 (0:00:00.023)       0:00:00.023 **** 
ok: [managed-node3]
META: ran handlers

TASK [Run the role] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:19
Wednesday 11 December 2024  10:07:26 -0500 (0:00:01.101)       0:00:01.125 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:07:26 -0500 (0:00:00.032)       0:00:01.157 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:07:26 -0500 (0:00:00.028)       0:00:01.186 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:07:26 -0500 (0:00:00.043)       0:00:01.229 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:07:27 -0500 (0:00:00.075)       0:00:01.305 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "exists": false
    }
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:07:27 -0500 (0:00:00.606)       0:00:01.911 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__storage_is_ostree": false
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:07:27 -0500 (0:00:00.100)       0:00:02.011 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:07:27 -0500 (0:00:00.039)       0:00:02.051 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:07:27 -0500 (0:00:00.033)       0:00:02.084 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:07:27 -0500 (0:00:00.106)       0:00:02.190 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:07:31 -0500 (0:00:03.379)       0:00:05.570 **** 
ok: [managed-node3] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:07:31 -0500 (0:00:00.066)       0:00:05.636 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:07:31 -0500 (0:00:00.064)       0:00:05.701 **** 
ok: [managed-node3] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:07:32 -0500 (0:00:00.798)       0:00:06.499 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Wednesday 11 December 2024  10:07:32 -0500 (0:00:00.100)       0:00:06.600 **** 

TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Wednesday 11 December 2024  10:07:32 -0500 (0:00:00.025)       0:00:06.625 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19
Wednesday 11 December 2024  10:07:32 -0500 (0:00:00.031)       0:00:06.657 **** 

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:07:32 -0500 (0:00:00.026)       0:00:06.684 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:07:35 -0500 (0:00:02.917)       0:00:09.601 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-hotplugd.service": {
                "name": "cloud-init-hotplugd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.portable1.service": {
                "name": "dbus-org.freedesktop.portable1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-makecache.service": {
                "name": "dnf-makecache.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dnf-system-upgrade-cleanup.service": {
                "name": "dnf-system-upgrade-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dnf-system-upgrade.service": {
                "name": "dnf-system-upgrade.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown-onfailure.service": {
                "name": "dracut-shutdown-onfailure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ebtables.service": {
                "name": "ebtables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "firewalld.service": {
                "name": "firewalld.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "grub-boot-indeterminate.service": {
                "name": "grub-boot-indeterminate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "import-state.service": {
                "name": "import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "iprdump.service": {
                "name": "iprdump.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprinit.service": {
                "name": "iprinit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "iprupdate.service": {
                "name": "iprupdate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "kvm_stat.service": {
                "name": "kvm_stat.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ldconfig.service": {
                "name": "ldconfig.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "loadmodules.service": {
                "name": "loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "lvm2-pvscan@8:0.service": {
                "name": "lvm2-pvscan@8:0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "man-db-cache-update.service": {
                "name": "man-db-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "man-db-restart-cache-update.service": {
                "name": "man-db-restart-cache-update.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "multipathd.service": {
                "name": "multipathd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "ndctl-monitor.service": {
                "name": "ndctl-monitor.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-convert.service": {
                "name": "nfs-convert.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfsdcld.service": {
                "name": "nfsdcld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nftables.service": {
                "name": "nftables.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nis-domainname.service": {
                "name": "nis-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "oddjobd.service": {
                "name": "oddjobd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "plymouth-halt.service": {
                "name": "plymouth-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-quit.service": {
                "name": "plymouth-quit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "plymouth-switch-root-initramfs.service": {
                "name": "plymouth-switch-root-initramfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "restraintd.service": {
                "name": "restraintd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rngd.service": {
                "name": "rngd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-autorelabel-mark.service": {
                "name": "selinux-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "selinux-autorelabel.service": {
                "name": "selinux-autorelabel.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@.service": {
                "name": "sshd-keygen@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "sshd-keygen@ecdsa.service": {
                "name": "sshd-keygen@ecdsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@ed25519.service": {
                "name": "sshd-keygen@ed25519.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd-keygen@rsa.service": {
                "name": "sshd-keygen@rsa.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "sssd-autofs.service": {
                "name": "sssd-autofs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-kcm.service": {
                "name": "sssd-kcm.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "sssd-nss.service": {
                "name": "sssd-nss.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pac.service": {
                "name": "sssd-pac.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-pam.service": {
                "name": "sssd-pam.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-ssh.service": {
                "name": "sssd-ssh.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd-sudo.service": {
                "name": "sssd-sudo.service",
                "source": "systemd",
                "state": "inactive",
                "status": "indirect"
            },
            "sssd.service": {
                "name": "sssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "stratis-fstab-setup@.service": {
                "name": "stratis-fstab-setup@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "stratisd-min-postinitrd.service": {
                "name": "stratisd-min-postinitrd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "stratisd.service": {
                "name": "stratisd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "active",
                "status": "enabled"
            },
            "system-update-cleanup.service": {
                "name": "system-update-cleanup.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-coredump@.service": {
                "name": "systemd-coredump@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-exit.service": {
                "name": "systemd-exit.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-portabled.service": {
                "name": "systemd-portabled.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-pstore.service": {
                "name": "systemd-pstore.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-resolved.service": {
                "name": "systemd-resolved.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-rfkill.service": {
                "name": "systemd-rfkill.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend-then-hibernate.service": {
                "name": "systemd-suspend-then-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "masked"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-volatile-root.service": {
                "name": "systemd-volatile-root.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "target.service": {
                "name": "target.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "targetclid.service": {
                "name": "targetclid.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "tcsd.service": {
                "name": "tcsd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "timedatex.service": {
                "name": "timedatex.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "unbound-anchor.service": {
                "name": "unbound-anchor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "user-runtime-dir@.service": {
                "name": "user-runtime-dir@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user-runtime-dir@0.service": {
                "name": "user-runtime-dir@0.service",
                "source": "systemd",
                "state": "stopped",
                "status": "unknown"
            },
            "user@.service": {
                "name": "user@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "user@0.service": {
                "name": "user@0.service",
                "source": "systemd",
                "state": "running",
                "status": "unknown"
            },
            "vdo-start-by-dev@.service": {
                "name": "vdo-start-by-dev@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "vdo.service": {
                "name": "vdo.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:07:37 -0500 (0:00:01.765)       0:00:11.366 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:07:37 -0500 (0:00:00.033)       0:00:11.400 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:07:37 -0500 (0:00:00.019)       0:00:11.419 **** 
ok: [managed-node3] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:07:37 -0500 (0:00:00.537)       0:00:11.957 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:07:37 -0500 (0:00:00.020)       0:00:11.977 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929635.1587687,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "ab8070345774adad92683e9645714452be7be474",
        "ctime": 1733929634.5777702,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929634.5777702,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1343,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:07:38 -0500 (0:00:00.382)       0:00:12.359 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:07:38 -0500 (0:00:00.023)       0:00:12.383 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:07:38 -0500 (0:00:00.047)       0:00:12.431 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [],
        "changed": false,
        "crypts": [],
        "failed": false,
        "leaves": [],
        "mounts": [],
        "packages": [],
        "pools": [],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:07:38 -0500 (0:00:00.028)       0:00:12.459 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:07:38 -0500 (0:00:00.022)       0:00:12.482 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:07:38 -0500 (0:00:00.023)       0:00:12.506 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:07:38 -0500 (0:00:00.019)       0:00:12.525 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:07:38 -0500 (0:00:00.019)       0:00:12.545 **** 

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:07:38 -0500 (0:00:00.020)       0:00:12.566 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:07:38 -0500 (0:00:00.019)       0:00:12.585 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:07:38 -0500 (0:00:00.020)       0:00:12.606 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:07:38 -0500 (0:00:00.393)       0:00:12.999 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:07:38 -0500 (0:00:00.019)       0:00:13.019 **** 
ok: [managed-node3]

TASK [Mark tasks to be skipped] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:23
Wednesday 11 December 2024  10:07:39 -0500 (0:00:00.788)       0:00:13.807 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_skip_checks": [
            "blivet_available",
            "packages_installed",
            "service_facts"
        ]
    },
    "changed": false
}

TASK [Get unused disks] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:30
Wednesday 11 December 2024  10:07:39 -0500 (0:00:00.025)       0:00:13.833 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml for managed-node3

TASK [Ensure test packages] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:2
Wednesday 11 December 2024  10:07:39 -0500 (0:00:00.031)       0:00:13.864 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Find unused disks in the system] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:11
Wednesday 11 December 2024  10:07:42 -0500 (0:00:02.859)       0:00:16.724 **** 
ok: [managed-node3] => {
    "changed": false,
    "disks": [
        "sda"
    ],
    "info": [
        "Line: NAME=\"/dev/sda\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/sdb\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/sdc\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/sdd\" TYPE=\"disk\" SIZE=\"1099511627776\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/sde\" TYPE=\"disk\" SIZE=\"1099511627776\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/sdf\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/sdg\" TYPE=\"disk\" SIZE=\"1099511627776\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/sdh\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/sdi\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/xvda\" TYPE=\"disk\" SIZE=\"268435456000\" FSTYPE=\"\" LOG-SEC=\"512\"",
        "Line: NAME=\"/dev/xvda1\" TYPE=\"part\" SIZE=\"268434390528\" FSTYPE=\"xfs\" LOG-SEC=\"512\"",
        "Line type [part] is not disk: NAME=\"/dev/xvda1\" TYPE=\"part\" SIZE=\"268434390528\" FSTYPE=\"xfs\" LOG-SEC=\"512\"",
        "filename [xvda1] is a partition",
        "Disk [/dev/xvda] attrs [{'type': 'disk', 'size': '268435456000', 'fstype': '', 'ssize': '512'}] has partitions"
    ]
}

TASK [Debug why there are no unused disks] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:20
Wednesday 11 December 2024  10:07:42 -0500 (0:00:00.567)       0:00:17.292 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set unused_disks if necessary] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:29
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.030)       0:00:17.322 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "unused_disks": [
            "sda"
        ]
    },
    "changed": false
}

TASK [Exit playbook when there's not enough unused disks in the system] ********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:34
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.042)       0:00:17.365 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Print unused disks] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:39
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.034)       0:00:17.399 **** 
ok: [managed-node3] => {
    "unused_disks": [
        "sda"
    ]
}

TASK [Create one LVM logical volume under one volume group with size 5g] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:38
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.038)       0:00:17.438 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.064)       0:00:17.502 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.045)       0:00:17.547 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.036)       0:00:17.583 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.078)       0:00:17.662 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.030)       0:00:17.693 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.029)       0:00:17.723 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.030)       0:00:17.754 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.031)       0:00:17.785 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.110)       0:00:17.896 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.032)       0:00:17.928 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "5g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.037)       0:00:17.966 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.054)       0:00:18.020 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.031)       0:00:18.051 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.031)       0:00:18.083 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.034)       0:00:18.117 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.033)       0:00:18.151 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.042)       0:00:18.193 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:07:43 -0500 (0:00:00.021)       0:00:18.215 **** 
changed: [managed-node3] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/sda",
            "fs_type": "lvmpv"
        },
        {
            "action": "create device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "create device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1",
        "/dev/mapper/foo-test1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs",
        "lvm2"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:07:48 -0500 (0:00:04.588)       0:00:22.803 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:07:48 -0500 (0:00:00.031)       0:00:22.834 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929635.1587687,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "ab8070345774adad92683e9645714452be7be474",
        "ctime": 1733929634.5777702,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929634.5777702,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1343,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:07:48 -0500 (0:00:00.460)       0:00:23.295 **** 
ok: [managed-node3] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:07:49 -0500 (0:00:00.530)       0:00:23.825 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:07:49 -0500 (0:00:00.020)       0:00:23.845 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/sda",
                "fs_type": "lvmpv"
            },
            {
                "action": "create device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "create device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1",
            "/dev/mapper/foo-test1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs",
            "lvm2"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:07:49 -0500 (0:00:00.027)       0:00:23.873 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:07:49 -0500 (0:00:00.032)       0:00:23.905 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:07:49 -0500 (0:00:00.031)       0:00:23.937 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:07:49 -0500 (0:00:00.027)       0:00:23.964 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:07:50 -0500 (0:00:00.931)       0:00:24.895 **** 
changed: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:07:51 -0500 (0:00:00.555)       0:00:25.451 **** 
skipping: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:07:51 -0500 (0:00:00.028)       0:00:25.479 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:07:51 -0500 (0:00:00.645)       0:00:26.125 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:07:52 -0500 (0:00:00.421)       0:00:26.546 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:07:52 -0500 (0:00:00.024)       0:00:26.571 **** 
ok: [managed-node3]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:55
Wednesday 11 December 2024  10:07:53 -0500 (0:00:00.821)       0:00:27.393 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node3

TASK [Print out pool information] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Wednesday 11 December 2024  10:07:53 -0500 (0:00:00.082)       0:00:27.475 **** 
ok: [managed-node3] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Wednesday 11 December 2024  10:07:53 -0500 (0:00:00.049)       0:00:27.525 **** 
skipping: [managed-node3] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Wednesday 11 December 2024  10:07:53 -0500 (0:00:00.046)       0:00:27.571 **** 
ok: [managed-node3] => {
    "changed": false,
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "ext4",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/foo-test1",
            "size": "5G",
            "type": "lvm",
            "uuid": "73110080-6417-4761-95bc-c7615e6d9219"
        },
        "/dev/sda": {
            "fstype": "LVM2_member",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "cpxLNK-7Ctf-10hw-mDam-h4iW-ZgCe-vIkbLe"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Wednesday 11 December 2024  10:07:54 -0500 (0:00:00.727)       0:00:28.298 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002708",
    "end": "2024-12-11 10:07:54.484720",
    "rc": 0,
    "start": "2024-12-11 10:07:54.482012"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Wednesday 11 December 2024  10:07:54 -0500 (0:00:00.540)       0:00:28.839 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002622",
    "end": "2024-12-11 10:07:54.881794",
    "failed_when_result": false,
    "rc": 0,
    "start": "2024-12-11 10:07:54.879172"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Wednesday 11 December 2024  10:07:54 -0500 (0:00:00.415)       0:00:29.254 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node3

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Wednesday 11 December 2024  10:07:55 -0500 (0:00:00.063)       0:00:29.317 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Wednesday 11 December 2024  10:07:55 -0500 (0:00:00.036)       0:00:29.354 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "vgs",
        "--noheadings",
        "--binary",
        "-o",
        "shared",
        "foo"
    ],
    "delta": "0:00:00.023567",
    "end": "2024-12-11 10:07:55.411362",
    "rc": 0,
    "start": "2024-12-11 10:07:55.387795"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Wednesday 11 December 2024  10:07:55 -0500 (0:00:00.415)       0:00:29.769 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Wednesday 11 December 2024  10:07:55 -0500 (0:00:00.040)       0:00:29.810 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Wednesday 11 December 2024  10:07:55 -0500 (0:00:00.072)       0:00:29.883 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Wednesday 11 December 2024  10:07:55 -0500 (0:00:00.041)       0:00:29.925 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sda",
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Wednesday 11 December 2024  10:07:56 -0500 (0:00:00.595)       0:00:30.520 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Wednesday 11 December 2024  10:07:56 -0500 (0:00:00.048)       0:00:30.569 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Wednesday 11 December 2024  10:07:56 -0500 (0:00:00.055)       0:00:30.625 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Wednesday 11 December 2024  10:07:56 -0500 (0:00:00.053)       0:00:30.679 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Wednesday 11 December 2024  10:07:56 -0500 (0:00:00.040)       0:00:30.720 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Wednesday 11 December 2024  10:07:56 -0500 (0:00:00.042)       0:00:30.762 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Wednesday 11 December 2024  10:07:56 -0500 (0:00:00.035)       0:00:30.798 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Wednesday 11 December 2024  10:07:56 -0500 (0:00:00.049)       0:00:30.847 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.41.57 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.713)       0:00:31.561 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.049)       0:00:31.610 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node3

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.084)       0:00:31.694 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.039)       0:00:31.734 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.034)       0:00:31.768 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.028)       0:00:31.797 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.036)       0:00:31.833 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.031)       0:00:31.865 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.033)       0:00:31.898 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.034)       0:00:31.933 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.039)       0:00:31.972 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.033)       0:00:32.005 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.033)       0:00:32.039 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.032)       0:00:32.071 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node3

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.070)       0:00:32.142 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node3

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.066)       0:00:32.209 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.030)       0:00:32.240 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Wednesday 11 December 2024  10:07:57 -0500 (0:00:00.034)       0:00:32.274 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.070)       0:00:32.344 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.033)       0:00:32.378 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.033)       0:00:32.412 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.032)       0:00:32.444 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.030)       0:00:32.474 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node3

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.070)       0:00:32.544 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node3

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.072)       0:00:32.616 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.042)       0:00:32.659 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.035)       0:00:32.695 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.032)       0:00:32.728 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.030)       0:00:32.758 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.075)       0:00:32.833 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.039)       0:00:32.873 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.037)       0:00:32.910 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node3

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.059)       0:00:32.969 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.036)       0:00:33.006 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.040)       0:00:33.046 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.035)       0:00:33.081 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.033)       0:00:33.115 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.031)       0:00:33.146 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.030)       0:00:33.176 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.030)       0:00:33.207 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node3

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Wednesday 11 December 2024  10:07:58 -0500 (0:00:00.073)       0:00:33.280 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node3

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.069)       0:00:33.350 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.033)       0:00:33.384 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.036)       0:00:33.420 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.033)       0:00:33.454 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.033)       0:00:33.488 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.035)       0:00:33.523 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.034)       0:00:33.558 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.031)       0:00:33.589 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node3

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.087)       0:00:33.676 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.041)       0:00:33.717 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.033)       0:00:33.751 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.032)       0:00:33.784 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.031)       0:00:33.816 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.031)       0:00:33.847 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.029)       0:00:33.877 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.034)       0:00:33.911 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node3

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.062)       0:00:33.974 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.038)       0:00:34.013 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node3

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.177)       0:00:34.190 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Wednesday 11 December 2024  10:07:59 -0500 (0:00:00.086)       0:00:34.276 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.042)       0:00:34.319 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.037)       0:00:34.356 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.048)       0:00:34.404 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.049)       0:00:34.454 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.078)       0:00:34.533 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.078)       0:00:34.612 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.062)       0:00:34.674 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.037)       0:00:34.712 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.042)       0:00:34.754 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.062)       0:00:34.817 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.128)       0:00:34.945 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.074)       0:00:35.020 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.058)       0:00:35.079 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.031)       0:00:35.110 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.047)       0:00:35.158 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.032)       0:00:35.190 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.045)       0:00:35.236 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Wednesday 11 December 2024  10:08:00 -0500 (0:00:00.055)       0:00:35.291 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929668.3846786,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1733929668.3846786,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 287887,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1733929668.3846786,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Wednesday 11 December 2024  10:08:01 -0500 (0:00:00.577)       0:00:35.869 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Wednesday 11 December 2024  10:08:01 -0500 (0:00:00.034)       0:00:35.904 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Wednesday 11 December 2024  10:08:01 -0500 (0:00:00.032)       0:00:35.937 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Wednesday 11 December 2024  10:08:01 -0500 (0:00:00.034)       0:00:35.971 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Wednesday 11 December 2024  10:08:01 -0500 (0:00:00.029)       0:00:36.001 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Wednesday 11 December 2024  10:08:01 -0500 (0:00:00.035)       0:00:36.037 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Wednesday 11 December 2024  10:08:01 -0500 (0:00:00.033)       0:00:36.070 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Wednesday 11 December 2024  10:08:01 -0500 (0:00:00.023)       0:00:36.094 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Wednesday 11 December 2024  10:08:04 -0500 (0:00:02.877)       0:00:38.971 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Wednesday 11 December 2024  10:08:04 -0500 (0:00:00.022)       0:00:38.993 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Wednesday 11 December 2024  10:08:04 -0500 (0:00:00.024)       0:00:39.018 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Wednesday 11 December 2024  10:08:04 -0500 (0:00:00.031)       0:00:39.049 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Wednesday 11 December 2024  10:08:04 -0500 (0:00:00.021)       0:00:39.071 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Wednesday 11 December 2024  10:08:04 -0500 (0:00:00.022)       0:00:39.093 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Wednesday 11 December 2024  10:08:04 -0500 (0:00:00.021)       0:00:39.114 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Wednesday 11 December 2024  10:08:04 -0500 (0:00:00.020)       0:00:39.135 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Wednesday 11 December 2024  10:08:04 -0500 (0:00:00.021)       0:00:39.157 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Wednesday 11 December 2024  10:08:04 -0500 (0:00:00.027)       0:00:39.184 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Wednesday 11 December 2024  10:08:04 -0500 (0:00:00.025)       0:00:39.210 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Wednesday 11 December 2024  10:08:04 -0500 (0:00:00.024)       0:00:39.234 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Wednesday 11 December 2024  10:08:04 -0500 (0:00:00.022)       0:00:39.257 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Wednesday 11 December 2024  10:08:04 -0500 (0:00:00.023)       0:00:39.280 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Wednesday 11 December 2024  10:08:05 -0500 (0:00:00.022)       0:00:39.302 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Wednesday 11 December 2024  10:08:05 -0500 (0:00:00.023)       0:00:39.325 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Wednesday 11 December 2024  10:08:05 -0500 (0:00:00.022)       0:00:39.348 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Wednesday 11 December 2024  10:08:05 -0500 (0:00:00.024)       0:00:39.373 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Wednesday 11 December 2024  10:08:05 -0500 (0:00:00.023)       0:00:39.396 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Wednesday 11 December 2024  10:08:05 -0500 (0:00:00.023)       0:00:39.419 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Wednesday 11 December 2024  10:08:05 -0500 (0:00:00.022)       0:00:39.441 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Wednesday 11 December 2024  10:08:05 -0500 (0:00:00.023)       0:00:39.464 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Wednesday 11 December 2024  10:08:05 -0500 (0:00:00.021)       0:00:39.485 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Wednesday 11 December 2024  10:08:05 -0500 (0:00:00.024)       0:00:39.510 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Wednesday 11 December 2024  10:08:05 -0500 (0:00:00.022)       0:00:39.533 **** 
ok: [managed-node3] => {
    "bytes": 5368709120,
    "changed": false,
    "lvm": "5g",
    "parted": "5GiB",
    "size": "5 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Wednesday 11 December 2024  10:08:05 -0500 (0:00:00.496)       0:00:40.030 **** 
ok: [managed-node3] => {
    "bytes": 5368709120,
    "changed": false,
    "lvm": "5g",
    "parted": "5GiB",
    "size": "5 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.373)       0:00:40.403 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_expected_size": "5368709120"
    },
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.030)       0:00:40.433 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.030)       0:00:40.464 **** 
ok: [managed-node3] => {
    "bytes": 10726680821,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.397)       0:00:40.862 **** 
skipping: [managed-node3] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.025)       0:00:40.888 **** 
skipping: [managed-node3] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.024)       0:00:40.912 **** 
skipping: [managed-node3] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.027)       0:00:40.940 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.034)       0:00:40.974 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.022)       0:00:40.997 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.054)       0:00:41.051 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.026)       0:00:41.078 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.023)       0:00:41.102 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.031)       0:00:41.134 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.042)       0:00:41.176 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.030)       0:00:41.207 **** 
skipping: [managed-node3] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.041)       0:00:41.248 **** 
skipping: [managed-node3] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Wednesday 11 December 2024  10:08:06 -0500 (0:00:00.046)       0:00:41.295 **** 
skipping: [managed-node3] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Wednesday 11 December 2024  10:08:07 -0500 (0:00:00.030)       0:00:41.325 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Wednesday 11 December 2024  10:08:07 -0500 (0:00:00.045)       0:00:41.371 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Wednesday 11 December 2024  10:08:07 -0500 (0:00:00.046)       0:00:41.417 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Wednesday 11 December 2024  10:08:07 -0500 (0:00:00.034)       0:00:41.451 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Wednesday 11 December 2024  10:08:07 -0500 (0:00:00.040)       0:00:41.492 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Wednesday 11 December 2024  10:08:07 -0500 (0:00:00.038)       0:00:41.530 **** 
ok: [managed-node3] => {
    "storage_test_actual_size": {
        "bytes": 5368709120,
        "changed": false,
        "failed": false,
        "lvm": "5g",
        "parted": "5GiB",
        "size": "5 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Wednesday 11 December 2024  10:08:07 -0500 (0:00:00.037)       0:00:41.568 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Wednesday 11 December 2024  10:08:07 -0500 (0:00:00.041)       0:00:41.609 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Wednesday 11 December 2024  10:08:07 -0500 (0:00:00.066)       0:00:41.676 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.023676",
    "end": "2024-12-11 10:08:07.913852",
    "rc": 0,
    "start": "2024-12-11 10:08:07.890176"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Wednesday 11 December 2024  10:08:07 -0500 (0:00:00.597)       0:00:42.274 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.036)       0:00:42.311 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.028)       0:00:42.339 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.029)       0:00:42.369 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.024)       0:00:42.394 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.030)       0:00:42.424 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.023)       0:00:42.447 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.026)       0:00:42.474 **** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.020)       0:00:42.494 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Change volume_size to 9g] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:58
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.030)       0:00:42.525 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.061)       0:00:42.587 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.034)       0:00:42.621 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.039)       0:00:42.661 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.089)       0:00:42.751 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.024)       0:00:42.775 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.031)       0:00:42.807 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.019)       0:00:42.827 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.028)       0:00:42.855 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.056)       0:00:42.912 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.028)       0:00:42.940 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "9g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.026)       0:00:42.967 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.023)       0:00:42.990 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.023)       0:00:43.013 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.022)       0:00:43.036 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.021)       0:00:43.057 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.033)       0:00:43.090 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.035)       0:00:43.126 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:08:08 -0500 (0:00:00.021)       0:00:43.147 **** 
changed: [managed-node3] => {
    "actions": [
        {
            "action": "resize device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "resize format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/mapper/foo-test1",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "e2fsprogs",
        "xfsprogs",
        "lvm2"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "9g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:08:13 -0500 (0:00:04.359)       0:00:47.507 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:08:13 -0500 (0:00:00.026)       0:00:47.534 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929671.0876713,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "306e713ef45e6db45ca0244c07767212954e6abb",
        "ctime": 1733929671.0846713,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929671.0846713,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1394,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:08:13 -0500 (0:00:00.375)       0:00:47.910 **** 
ok: [managed-node3] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:08:14 -0500 (0:00:00.455)       0:00:48.366 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:08:14 -0500 (0:00:00.035)       0:00:48.401 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [
            {
                "action": "resize device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "resize format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/mapper/foo-test1",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "e2fsprogs",
            "xfsprogs",
            "lvm2"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "9g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:08:14 -0500 (0:00:00.043)       0:00:48.444 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "9g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:08:14 -0500 (0:00:00.048)       0:00:48.493 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:08:14 -0500 (0:00:00.041)       0:00:48.535 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:08:14 -0500 (0:00:00.032)       0:00:48.567 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:08:14 -0500 (0:00:00.695)       0:00:49.263 **** 
ok: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:08:15 -0500 (0:00:00.423)       0:00:49.686 **** 
skipping: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:08:15 -0500 (0:00:00.046)       0:00:49.732 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:08:16 -0500 (0:00:00.671)       0:00:50.404 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:08:16 -0500 (0:00:00.372)       0:00:50.776 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:08:16 -0500 (0:00:00.019)       0:00:50.796 **** 
ok: [managed-node3]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:72
Wednesday 11 December 2024  10:08:17 -0500 (0:00:00.763)       0:00:51.560 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node3

TASK [Print out pool information] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Wednesday 11 December 2024  10:08:17 -0500 (0:00:00.036)       0:00:51.596 **** 
ok: [managed-node3] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "9g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Wednesday 11 December 2024  10:08:17 -0500 (0:00:00.029)       0:00:51.626 **** 
skipping: [managed-node3] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Wednesday 11 December 2024  10:08:17 -0500 (0:00:00.022)       0:00:51.648 **** 
ok: [managed-node3] => {
    "changed": false,
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "ext4",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/foo-test1",
            "size": "9G",
            "type": "lvm",
            "uuid": "73110080-6417-4761-95bc-c7615e6d9219"
        },
        "/dev/sda": {
            "fstype": "LVM2_member",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "cpxLNK-7Ctf-10hw-mDam-h4iW-ZgCe-vIkbLe"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Wednesday 11 December 2024  10:08:17 -0500 (0:00:00.371)       0:00:52.019 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003187",
    "end": "2024-12-11 10:08:18.052325",
    "rc": 0,
    "start": "2024-12-11 10:08:18.049138"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Wednesday 11 December 2024  10:08:18 -0500 (0:00:00.388)       0:00:52.408 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003325",
    "end": "2024-12-11 10:08:18.434223",
    "failed_when_result": false,
    "rc": 0,
    "start": "2024-12-11 10:08:18.430898"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Wednesday 11 December 2024  10:08:18 -0500 (0:00:00.390)       0:00:52.799 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node3

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Wednesday 11 December 2024  10:08:18 -0500 (0:00:00.051)       0:00:52.850 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Wednesday 11 December 2024  10:08:18 -0500 (0:00:00.021)       0:00:52.871 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "vgs",
        "--noheadings",
        "--binary",
        "-o",
        "shared",
        "foo"
    ],
    "delta": "0:00:00.025097",
    "end": "2024-12-11 10:08:18.923011",
    "rc": 0,
    "start": "2024-12-11 10:08:18.897914"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Wednesday 11 December 2024  10:08:18 -0500 (0:00:00.416)       0:00:53.288 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Wednesday 11 December 2024  10:08:19 -0500 (0:00:00.070)       0:00:53.359 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Wednesday 11 December 2024  10:08:19 -0500 (0:00:00.097)       0:00:53.456 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Wednesday 11 December 2024  10:08:19 -0500 (0:00:00.110)       0:00:53.567 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sda",
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Wednesday 11 December 2024  10:08:19 -0500 (0:00:00.464)       0:00:54.032 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Wednesday 11 December 2024  10:08:19 -0500 (0:00:00.037)       0:00:54.069 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Wednesday 11 December 2024  10:08:19 -0500 (0:00:00.038)       0:00:54.108 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Wednesday 11 December 2024  10:08:19 -0500 (0:00:00.039)       0:00:54.148 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Wednesday 11 December 2024  10:08:19 -0500 (0:00:00.043)       0:00:54.192 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Wednesday 11 December 2024  10:08:19 -0500 (0:00:00.042)       0:00:54.235 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Wednesday 11 December 2024  10:08:20 -0500 (0:00:00.069)       0:00:54.304 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Wednesday 11 December 2024  10:08:20 -0500 (0:00:00.046)       0:00:54.351 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.41.57 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Wednesday 11 December 2024  10:08:20 -0500 (0:00:00.485)       0:00:54.837 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Wednesday 11 December 2024  10:08:20 -0500 (0:00:00.037)       0:00:54.874 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node3

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Wednesday 11 December 2024  10:08:20 -0500 (0:00:00.070)       0:00:54.945 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Wednesday 11 December 2024  10:08:20 -0500 (0:00:00.041)       0:00:54.986 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Wednesday 11 December 2024  10:08:20 -0500 (0:00:00.032)       0:00:55.018 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Wednesday 11 December 2024  10:08:20 -0500 (0:00:00.022)       0:00:55.041 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Wednesday 11 December 2024  10:08:20 -0500 (0:00:00.035)       0:00:55.076 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Wednesday 11 December 2024  10:08:20 -0500 (0:00:00.027)       0:00:55.104 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Wednesday 11 December 2024  10:08:20 -0500 (0:00:00.029)       0:00:55.133 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Wednesday 11 December 2024  10:08:20 -0500 (0:00:00.031)       0:00:55.165 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Wednesday 11 December 2024  10:08:20 -0500 (0:00:00.024)       0:00:55.189 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Wednesday 11 December 2024  10:08:20 -0500 (0:00:00.026)       0:00:55.216 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Wednesday 11 December 2024  10:08:20 -0500 (0:00:00.026)       0:00:55.243 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Wednesday 11 December 2024  10:08:20 -0500 (0:00:00.025)       0:00:55.268 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node3

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.042)       0:00:55.311 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node3

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.044)       0:00:55.355 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.021)       0:00:55.377 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.021)       0:00:55.398 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.021)       0:00:55.419 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.021)       0:00:55.441 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.039)       0:00:55.481 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.026)       0:00:55.508 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.029)       0:00:55.537 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node3

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.067)       0:00:55.605 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node3

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.068)       0:00:55.673 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.035)       0:00:55.708 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.035)       0:00:55.744 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.026)       0:00:55.771 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.026)       0:00:55.797 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.108)       0:00:55.906 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.028)       0:00:55.934 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.031)       0:00:55.966 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node3

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.043)       0:00:56.010 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.026)       0:00:56.036 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.027)       0:00:56.064 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.026)       0:00:56.090 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.024)       0:00:56.115 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.036)       0:00:56.152 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.034)       0:00:56.186 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.030)       0:00:56.217 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node3

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Wednesday 11 December 2024  10:08:21 -0500 (0:00:00.073)       0:00:56.290 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node3

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.085)       0:00:56.376 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.037)       0:00:56.414 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.034)       0:00:56.448 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.043)       0:00:56.492 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.047)       0:00:56.539 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.041)       0:00:56.581 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.037)       0:00:56.619 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.034)       0:00:56.653 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node3

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.084)       0:00:56.738 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.033)       0:00:56.771 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.034)       0:00:56.805 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.032)       0:00:56.837 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.033)       0:00:56.870 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.031)       0:00:56.902 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.033)       0:00:56.936 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.030)       0:00:56.966 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node3

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.059)       0:00:57.026 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.048)       0:00:57.074 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node3

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.180)       0:00:57.255 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Wednesday 11 December 2024  10:08:22 -0500 (0:00:00.038)       0:00:57.293 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.043)       0:00:57.337 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.046)       0:00:57.383 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.037)       0:00:57.420 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.027)       0:00:57.448 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.028)       0:00:57.476 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.044)       0:00:57.521 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.033)       0:00:57.554 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.035)       0:00:57.590 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.051)       0:00:57.642 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.035)       0:00:57.677 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.049)       0:00:57.727 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.027)       0:00:57.754 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.026)       0:00:57.780 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.021)       0:00:57.802 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.027)       0:00:57.830 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.058)       0:00:57.889 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.039)       0:00:57.928 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Wednesday 11 December 2024  10:08:23 -0500 (0:00:00.064)       0:00:57.992 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929693.0646117,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1733929693.0636117,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 287887,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1733929693.0636117,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Wednesday 11 December 2024  10:08:24 -0500 (0:00:00.582)       0:00:58.575 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Wednesday 11 December 2024  10:08:24 -0500 (0:00:00.052)       0:00:58.628 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Wednesday 11 December 2024  10:08:24 -0500 (0:00:00.049)       0:00:58.677 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Wednesday 11 December 2024  10:08:24 -0500 (0:00:00.060)       0:00:58.738 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Wednesday 11 December 2024  10:08:24 -0500 (0:00:00.049)       0:00:58.787 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Wednesday 11 December 2024  10:08:24 -0500 (0:00:00.052)       0:00:58.840 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Wednesday 11 December 2024  10:08:24 -0500 (0:00:00.049)       0:00:58.889 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Wednesday 11 December 2024  10:08:24 -0500 (0:00:00.034)       0:00:58.923 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Wednesday 11 December 2024  10:08:27 -0500 (0:00:03.060)       0:01:01.984 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Wednesday 11 December 2024  10:08:27 -0500 (0:00:00.032)       0:01:02.016 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Wednesday 11 December 2024  10:08:27 -0500 (0:00:00.031)       0:01:02.048 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Wednesday 11 December 2024  10:08:27 -0500 (0:00:00.045)       0:01:02.094 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Wednesday 11 December 2024  10:08:27 -0500 (0:00:00.040)       0:01:02.134 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Wednesday 11 December 2024  10:08:27 -0500 (0:00:00.031)       0:01:02.165 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Wednesday 11 December 2024  10:08:27 -0500 (0:00:00.032)       0:01:02.198 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Wednesday 11 December 2024  10:08:27 -0500 (0:00:00.027)       0:01:02.225 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Wednesday 11 December 2024  10:08:27 -0500 (0:00:00.034)       0:01:02.260 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.041)       0:01:02.301 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.038)       0:01:02.340 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.029)       0:01:02.369 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.031)       0:01:02.401 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.032)       0:01:02.433 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.033)       0:01:02.467 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.031)       0:01:02.498 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.031)       0:01:02.530 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.032)       0:01:02.562 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.031)       0:01:02.594 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.034)       0:01:02.628 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.035)       0:01:02.664 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.035)       0:01:02.700 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.033)       0:01:02.733 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.032)       0:01:02.766 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.032)       0:01:02.798 **** 
ok: [managed-node3] => {
    "bytes": 9663676416,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Wednesday 11 December 2024  10:08:28 -0500 (0:00:00.469)       0:01:03.268 **** 
ok: [managed-node3] => {
    "bytes": 9663676416,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Wednesday 11 December 2024  10:08:29 -0500 (0:00:00.442)       0:01:03.711 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_expected_size": "9663676416"
    },
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Wednesday 11 December 2024  10:08:29 -0500 (0:00:00.045)       0:01:03.756 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "9663676416"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Wednesday 11 December 2024  10:08:29 -0500 (0:00:00.032)       0:01:03.789 **** 
ok: [managed-node3] => {
    "bytes": 10726680821,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.547)       0:01:04.337 **** 
skipping: [managed-node3] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.040)       0:01:04.378 **** 
skipping: [managed-node3] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.036)       0:01:04.414 **** 
skipping: [managed-node3] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.029)       0:01:04.444 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.039)       0:01:04.483 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.029)       0:01:04.512 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.035)       0:01:04.548 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.024)       0:01:04.573 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.029)       0:01:04.602 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.032)       0:01:04.635 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.020)       0:01:04.655 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.023)       0:01:04.679 **** 
skipping: [managed-node3] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.020)       0:01:04.700 **** 
skipping: [managed-node3] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.021)       0:01:04.721 **** 
skipping: [managed-node3] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.021)       0:01:04.743 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.020)       0:01:04.763 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.023)       0:01:04.787 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.023)       0:01:04.811 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.068)       0:01:04.880 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.027)       0:01:04.907 **** 
ok: [managed-node3] => {
    "storage_test_actual_size": {
        "bytes": 9663676416,
        "changed": false,
        "failed": false,
        "lvm": "9g",
        "parted": "9GiB",
        "size": "9 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.041)       0:01:04.949 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "9663676416"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.026)       0:01:04.976 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Wednesday 11 December 2024  10:08:30 -0500 (0:00:00.039)       0:01:05.015 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.023846",
    "end": "2024-12-11 10:08:31.058931",
    "rc": 0,
    "start": "2024-12-11 10:08:31.035085"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.397)       0:01:05.413 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.033)       0:01:05.447 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.031)       0:01:05.478 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.026)       0:01:05.505 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.024)       0:01:05.530 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.022)       0:01:05.552 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.023)       0:01:05.576 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.023)       0:01:05.599 **** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.019)       0:01:05.618 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Change volume size to 5g] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:75
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.022)       0:01:05.641 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.048)       0:01:05.689 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.045)       0:01:05.735 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.039)       0:01:05.775 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.086)       0:01:05.861 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.032)       0:01:05.893 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.033)       0:01:05.927 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.033)       0:01:05.961 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.039)       0:01:06.001 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.076)       0:01:06.077 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.035)       0:01:06.113 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "5g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.038)       0:01:06.151 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.033)       0:01:06.184 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.032)       0:01:06.217 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.032)       0:01:06.249 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:08:31 -0500 (0:00:00.031)       0:01:06.281 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:08:32 -0500 (0:00:00.034)       0:01:06.315 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:08:32 -0500 (0:00:00.049)       0:01:06.365 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:08:32 -0500 (0:00:00.029)       0:01:06.394 **** 
changed: [managed-node3] => {
    "actions": [
        {
            "action": "resize format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        },
        {
            "action": "resize device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/mapper/foo-test1",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "e2fsprogs",
        "lvm2",
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:08:37 -0500 (0:00:04.957)       0:01:11.352 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:08:37 -0500 (0:00:00.022)       0:01:11.375 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929671.0876713,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "306e713ef45e6db45ca0244c07767212954e6abb",
        "ctime": 1733929671.0846713,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929671.0846713,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1394,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:08:37 -0500 (0:00:00.386)       0:01:11.761 **** 
ok: [managed-node3] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:08:37 -0500 (0:00:00.383)       0:01:12.144 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:08:37 -0500 (0:00:00.020)       0:01:12.164 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [
            {
                "action": "resize format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            },
            {
                "action": "resize device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/mapper/foo-test1",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "e2fsprogs",
            "lvm2",
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:08:37 -0500 (0:00:00.026)       0:01:12.191 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:08:37 -0500 (0:00:00.027)       0:01:12.218 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:08:37 -0500 (0:00:00.025)       0:01:12.243 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:08:37 -0500 (0:00:00.021)       0:01:12.265 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:08:38 -0500 (0:00:00.635)       0:01:12.901 **** 
changed: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:08:39 -0500 (0:00:00.406)       0:01:13.308 **** 
skipping: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:08:39 -0500 (0:00:00.028)       0:01:13.337 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:08:39 -0500 (0:00:00.645)       0:01:13.982 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:08:40 -0500 (0:00:00.373)       0:01:14.356 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:08:40 -0500 (0:00:00.018)       0:01:14.375 **** 
ok: [managed-node3]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:89
Wednesday 11 December 2024  10:08:40 -0500 (0:00:00.761)       0:01:15.136 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node3

TASK [Print out pool information] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Wednesday 11 December 2024  10:08:40 -0500 (0:00:00.058)       0:01:15.194 **** 
ok: [managed-node3] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Wednesday 11 December 2024  10:08:40 -0500 (0:00:00.032)       0:01:15.227 **** 
skipping: [managed-node3] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Wednesday 11 December 2024  10:08:40 -0500 (0:00:00.027)       0:01:15.255 **** 
ok: [managed-node3] => {
    "changed": false,
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "ext4",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/foo-test1",
            "size": "5G",
            "type": "lvm",
            "uuid": "73110080-6417-4761-95bc-c7615e6d9219"
        },
        "/dev/sda": {
            "fstype": "LVM2_member",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "cpxLNK-7Ctf-10hw-mDam-h4iW-ZgCe-vIkbLe"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Wednesday 11 December 2024  10:08:41 -0500 (0:00:00.391)       0:01:15.647 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002593",
    "end": "2024-12-11 10:08:41.689672",
    "rc": 0,
    "start": "2024-12-11 10:08:41.687079"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Wednesday 11 December 2024  10:08:41 -0500 (0:00:00.397)       0:01:16.044 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002557",
    "end": "2024-12-11 10:08:42.056303",
    "failed_when_result": false,
    "rc": 0,
    "start": "2024-12-11 10:08:42.053746"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Wednesday 11 December 2024  10:08:42 -0500 (0:00:00.379)       0:01:16.423 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node3

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Wednesday 11 December 2024  10:08:42 -0500 (0:00:00.049)       0:01:16.472 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Wednesday 11 December 2024  10:08:42 -0500 (0:00:00.026)       0:01:16.499 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "vgs",
        "--noheadings",
        "--binary",
        "-o",
        "shared",
        "foo"
    ],
    "delta": "0:00:00.023708",
    "end": "2024-12-11 10:08:42.539927",
    "rc": 0,
    "start": "2024-12-11 10:08:42.516219"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Wednesday 11 December 2024  10:08:42 -0500 (0:00:00.401)       0:01:16.900 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Wednesday 11 December 2024  10:08:42 -0500 (0:00:00.037)       0:01:16.937 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Wednesday 11 December 2024  10:08:42 -0500 (0:00:00.089)       0:01:17.027 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Wednesday 11 December 2024  10:08:42 -0500 (0:00:00.030)       0:01:17.057 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sda",
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Wednesday 11 December 2024  10:08:43 -0500 (0:00:00.374)       0:01:17.432 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Wednesday 11 December 2024  10:08:43 -0500 (0:00:00.028)       0:01:17.461 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Wednesday 11 December 2024  10:08:43 -0500 (0:00:00.040)       0:01:17.501 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Wednesday 11 December 2024  10:08:43 -0500 (0:00:00.036)       0:01:17.538 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Wednesday 11 December 2024  10:08:43 -0500 (0:00:00.034)       0:01:17.572 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Wednesday 11 December 2024  10:08:43 -0500 (0:00:00.031)       0:01:17.604 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Wednesday 11 December 2024  10:08:43 -0500 (0:00:00.025)       0:01:17.630 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Wednesday 11 December 2024  10:08:43 -0500 (0:00:00.037)       0:01:17.668 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.41.57 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Wednesday 11 December 2024  10:08:43 -0500 (0:00:00.416)       0:01:18.084 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Wednesday 11 December 2024  10:08:43 -0500 (0:00:00.028)       0:01:18.112 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node3

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Wednesday 11 December 2024  10:08:43 -0500 (0:00:00.042)       0:01:18.155 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Wednesday 11 December 2024  10:08:43 -0500 (0:00:00.022)       0:01:18.177 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Wednesday 11 December 2024  10:08:43 -0500 (0:00:00.022)       0:01:18.200 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Wednesday 11 December 2024  10:08:43 -0500 (0:00:00.023)       0:01:18.223 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Wednesday 11 December 2024  10:08:43 -0500 (0:00:00.023)       0:01:18.247 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Wednesday 11 December 2024  10:08:43 -0500 (0:00:00.031)       0:01:18.279 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.021)       0:01:18.301 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.021)       0:01:18.322 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.021)       0:01:18.344 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.023)       0:01:18.367 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.022)       0:01:18.389 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.021)       0:01:18.410 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node3

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.042)       0:01:18.453 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node3

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.046)       0:01:18.499 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.022)       0:01:18.522 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.022)       0:01:18.544 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.021)       0:01:18.566 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.026)       0:01:18.592 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.026)       0:01:18.619 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.029)       0:01:18.648 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.032)       0:01:18.681 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node3

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.068)       0:01:18.750 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node3

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.064)       0:01:18.814 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.025)       0:01:18.840 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.026)       0:01:18.867 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.023)       0:01:18.891 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.025)       0:01:18.916 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.053)       0:01:18.969 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.028)       0:01:18.997 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.026)       0:01:19.024 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node3

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.071)       0:01:19.096 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.026)       0:01:19.122 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.026)       0:01:19.148 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.023)       0:01:19.172 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.023)       0:01:19.196 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.028)       0:01:19.224 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.030)       0:01:19.255 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Wednesday 11 December 2024  10:08:44 -0500 (0:00:00.024)       0:01:19.279 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node3

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.068)       0:01:19.348 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node3

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.054)       0:01:19.402 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.027)       0:01:19.430 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.021)       0:01:19.451 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.022)       0:01:19.473 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.023)       0:01:19.497 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.021)       0:01:19.519 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.021)       0:01:19.540 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.020)       0:01:19.560 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node3

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.055)       0:01:19.616 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.024)       0:01:19.640 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.032)       0:01:19.672 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.027)       0:01:19.700 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.024)       0:01:19.725 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.025)       0:01:19.750 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.025)       0:01:19.776 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.023)       0:01:19.799 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node3

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.045)       0:01:19.844 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.028)       0:01:19.873 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node3

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.114)       0:01:19.987 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.026)       0:01:20.013 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.035)       0:01:20.048 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.023)       0:01:20.072 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.039)       0:01:20.111 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.029)       0:01:20.141 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.032)       0:01:20.173 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.029)       0:01:20.203 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.031)       0:01:20.234 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.022)       0:01:20.257 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Wednesday 11 December 2024  10:08:45 -0500 (0:00:00.027)       0:01:20.284 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.029)       0:01:20.313 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.061)       0:01:20.375 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.030)       0:01:20.405 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.024)       0:01:20.430 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.020)       0:01:20.451 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.029)       0:01:20.480 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.021)       0:01:20.501 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.029)       0:01:20.531 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.031)       0:01:20.562 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929716.9145448,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1733929716.9145448,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 287887,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1733929716.9145448,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.419)       0:01:20.981 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.028)       0:01:21.010 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.022)       0:01:21.033 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.027)       0:01:21.060 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.028)       0:01:21.088 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.039)       0:01:21.128 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.028)       0:01:21.157 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Wednesday 11 December 2024  10:08:46 -0500 (0:00:00.033)       0:01:21.190 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Wednesday 11 December 2024  10:08:49 -0500 (0:00:02.913)       0:01:24.104 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Wednesday 11 December 2024  10:08:49 -0500 (0:00:00.024)       0:01:24.128 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Wednesday 11 December 2024  10:08:49 -0500 (0:00:00.026)       0:01:24.154 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Wednesday 11 December 2024  10:08:49 -0500 (0:00:00.034)       0:01:24.188 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Wednesday 11 December 2024  10:08:49 -0500 (0:00:00.028)       0:01:24.217 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Wednesday 11 December 2024  10:08:49 -0500 (0:00:00.024)       0:01:24.241 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Wednesday 11 December 2024  10:08:49 -0500 (0:00:00.023)       0:01:24.265 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Wednesday 11 December 2024  10:08:49 -0500 (0:00:00.022)       0:01:24.288 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.022)       0:01:24.310 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.027)       0:01:24.338 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.026)       0:01:24.364 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.021)       0:01:24.386 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.027)       0:01:24.414 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.024)       0:01:24.439 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.030)       0:01:24.469 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.046)       0:01:24.516 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.042)       0:01:24.558 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.040)       0:01:24.599 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.037)       0:01:24.636 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.033)       0:01:24.669 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.036)       0:01:24.706 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.034)       0:01:24.740 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.037)       0:01:24.778 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.041)       0:01:24.820 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.037)       0:01:24.857 **** 
ok: [managed-node3] => {
    "bytes": 5368709120,
    "changed": false,
    "lvm": "5g",
    "parted": "5GiB",
    "size": "5 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Wednesday 11 December 2024  10:08:50 -0500 (0:00:00.390)       0:01:25.248 **** 
ok: [managed-node3] => {
    "bytes": 5368709120,
    "changed": false,
    "lvm": "5g",
    "parted": "5GiB",
    "size": "5 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Wednesday 11 December 2024  10:08:51 -0500 (0:00:00.364)       0:01:25.613 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_expected_size": "5368709120"
    },
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Wednesday 11 December 2024  10:08:51 -0500 (0:00:00.030)       0:01:25.643 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Wednesday 11 December 2024  10:08:51 -0500 (0:00:00.023)       0:01:25.666 **** 
ok: [managed-node3] => {
    "bytes": 10726680821,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Wednesday 11 December 2024  10:08:51 -0500 (0:00:00.369)       0:01:26.035 **** 
skipping: [managed-node3] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Wednesday 11 December 2024  10:08:51 -0500 (0:00:00.025)       0:01:26.061 **** 
skipping: [managed-node3] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Wednesday 11 December 2024  10:08:51 -0500 (0:00:00.025)       0:01:26.086 **** 
skipping: [managed-node3] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Wednesday 11 December 2024  10:08:51 -0500 (0:00:00.026)       0:01:26.112 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Wednesday 11 December 2024  10:08:51 -0500 (0:00:00.025)       0:01:26.138 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Wednesday 11 December 2024  10:08:51 -0500 (0:00:00.021)       0:01:26.160 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Wednesday 11 December 2024  10:08:51 -0500 (0:00:00.020)       0:01:26.180 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Wednesday 11 December 2024  10:08:51 -0500 (0:00:00.020)       0:01:26.201 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Wednesday 11 December 2024  10:08:51 -0500 (0:00:00.021)       0:01:26.223 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Wednesday 11 December 2024  10:08:51 -0500 (0:00:00.023)       0:01:26.246 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Wednesday 11 December 2024  10:08:51 -0500 (0:00:00.023)       0:01:26.269 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Wednesday 11 December 2024  10:08:51 -0500 (0:00:00.021)       0:01:26.291 **** 
skipping: [managed-node3] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.021)       0:01:26.313 **** 
skipping: [managed-node3] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.021)       0:01:26.334 **** 
skipping: [managed-node3] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.022)       0:01:26.356 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.022)       0:01:26.379 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.021)       0:01:26.400 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.021)       0:01:26.421 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.021)       0:01:26.443 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.022)       0:01:26.465 **** 
ok: [managed-node3] => {
    "storage_test_actual_size": {
        "bytes": 5368709120,
        "changed": false,
        "failed": false,
        "lvm": "5g",
        "parted": "5GiB",
        "size": "5 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.025)       0:01:26.491 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.024)       0:01:26.515 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.030)       0:01:26.546 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.030579",
    "end": "2024-12-11 10:08:52.608041",
    "rc": 0,
    "start": "2024-12-11 10:08:52.577462"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.415)       0:01:26.962 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.027)       0:01:26.990 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.029)       0:01:27.020 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.025)       0:01:27.045 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.058)       0:01:27.104 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.024)       0:01:27.128 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.023)       0:01:27.151 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.021)       0:01:27.172 **** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.018)       0:01:27.191 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Test for correct handling of too-large volume size] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:92
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.019)       0:01:27.211 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node3

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.042)       0:01:27.254 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_pools_global": [],
        "storage_safe_mode_global": false,
        "storage_volumes_global": []
    },
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Wednesday 11 December 2024  10:08:52 -0500 (0:00:00.026)       0:01:27.281 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.034)       0:01:27.316 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.030)       0:01:27.347 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.025)       0:01:27.372 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.055)       0:01:27.428 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.023)       0:01:27.451 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.021)       0:01:27.472 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.020)       0:01:27.492 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.021)       0:01:27.514 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.051)       0:01:27.566 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.025)       0:01:27.591 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "12884901888.0"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.039)       0:01:27.631 **** 
ok: [managed-node3] => {
    "storage_volumes": []
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.037)       0:01:27.669 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.023)       0:01:27.692 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.043)       0:01:27.735 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.034)       0:01:27.770 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.035)       0:01:27.805 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.052)       0:01:27.857 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:08:53 -0500 (0:00:00.031)       0:01:27.889 **** 
fatal: [managed-node3]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

volume 'test1' cannot be resized to '12 GiB'

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Wednesday 11 December 2024  10:08:57 -0500 (0:00:04.344)       0:01:32.233 **** 
fatal: [managed-node3]: FAILED! => {
    "changed": false
}

MSG:

{'msg': "volume 'test1' cannot be resized to '12 GiB'", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'encryption_clevis_pin': None, 'encryption_tang_url': None, 'encryption_tang_thumbprint': None, 'grow_to_fill': False, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'shared': False, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '12884901888.0', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}]}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'grow_to_fill': False, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, 'shared': False}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': False, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:08:57 -0500 (0:00:00.028)       0:01:32.262 **** 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Wednesday 11 December 2024  10:08:57 -0500 (0:00:00.019)       0:01:32.281 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.024)       0:01:32.306 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.031)       0:01:32.337 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Try to create LVM with volume size equal disk's size, resize to 10737418240] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:110
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.025)       0:01:32.362 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.033)       0:01:32.396 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.029)       0:01:32.426 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.025)       0:01:32.451 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.115)       0:01:32.567 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.023)       0:01:32.591 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.022)       0:01:32.613 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.020)       0:01:32.634 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.019)       0:01:32.653 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.051)       0:01:32.704 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.028)       0:01:32.733 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "10737418240"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.035)       0:01:32.769 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.023)       0:01:32.792 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.024)       0:01:32.816 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.022)       0:01:32.839 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.020)       0:01:32.860 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.021)       0:01:32.881 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.033)       0:01:32.915 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:08:58 -0500 (0:00:00.020)       0:01:32.936 **** 
changed: [managed-node3] => {
    "actions": [
        {
            "action": "resize device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "resize format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/mapper/foo-test1",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "lvm2",
        "e2fsprogs",
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "10737418240",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:09:03 -0500 (0:00:04.474)       0:01:37.410 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:09:03 -0500 (0:00:00.022)       0:01:37.433 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929671.0876713,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "306e713ef45e6db45ca0244c07767212954e6abb",
        "ctime": 1733929671.0846713,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929671.0846713,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1394,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:09:03 -0500 (0:00:00.385)       0:01:37.819 **** 
ok: [managed-node3] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:09:03 -0500 (0:00:00.389)       0:01:38.208 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:09:03 -0500 (0:00:00.020)       0:01:38.229 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [
            {
                "action": "resize device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "resize format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/mapper/foo-test1",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "lvm2",
            "e2fsprogs",
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "10737418240",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:09:03 -0500 (0:00:00.027)       0:01:38.256 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "10737418240",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:09:03 -0500 (0:00:00.025)       0:01:38.282 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:09:04 -0500 (0:00:00.024)       0:01:38.306 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:09:04 -0500 (0:00:00.021)       0:01:38.328 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:09:04 -0500 (0:00:00.636)       0:01:38.964 **** 
ok: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:09:05 -0500 (0:00:00.407)       0:01:39.372 **** 
skipping: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:09:05 -0500 (0:00:00.029)       0:01:39.402 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:09:05 -0500 (0:00:00.638)       0:01:40.040 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:09:06 -0500 (0:00:00.381)       0:01:40.422 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:09:06 -0500 (0:00:00.021)       0:01:40.443 **** 
ok: [managed-node3]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:125
Wednesday 11 December 2024  10:09:06 -0500 (0:00:00.816)       0:01:41.260 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node3

TASK [Print out pool information] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Wednesday 11 December 2024  10:09:07 -0500 (0:00:00.047)       0:01:41.307 **** 
ok: [managed-node3] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "10737418240",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Wednesday 11 December 2024  10:09:07 -0500 (0:00:00.189)       0:01:41.496 **** 
skipping: [managed-node3] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Wednesday 11 December 2024  10:09:07 -0500 (0:00:00.083)       0:01:41.580 **** 
ok: [managed-node3] => {
    "changed": false,
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "ext4",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/foo-test1",
            "size": "10G",
            "type": "lvm",
            "uuid": "73110080-6417-4761-95bc-c7615e6d9219"
        },
        "/dev/sda": {
            "fstype": "LVM2_member",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "cpxLNK-7Ctf-10hw-mDam-h4iW-ZgCe-vIkbLe"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Wednesday 11 December 2024  10:09:07 -0500 (0:00:00.441)       0:01:42.022 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002613",
    "end": "2024-12-11 10:09:08.125823",
    "rc": 0,
    "start": "2024-12-11 10:09:08.123210"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Wednesday 11 December 2024  10:09:08 -0500 (0:00:00.480)       0:01:42.502 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002468",
    "end": "2024-12-11 10:09:08.605833",
    "failed_when_result": false,
    "rc": 0,
    "start": "2024-12-11 10:09:08.603365"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Wednesday 11 December 2024  10:09:08 -0500 (0:00:00.474)       0:01:42.977 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node3

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Wednesday 11 December 2024  10:09:08 -0500 (0:00:00.120)       0:01:43.098 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Wednesday 11 December 2024  10:09:08 -0500 (0:00:00.029)       0:01:43.128 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "vgs",
        "--noheadings",
        "--binary",
        "-o",
        "shared",
        "foo"
    ],
    "delta": "0:00:00.107176",
    "end": "2024-12-11 10:09:09.356123",
    "rc": 0,
    "start": "2024-12-11 10:09:09.248947"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Wednesday 11 December 2024  10:09:09 -0500 (0:00:00.604)       0:01:43.732 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Wednesday 11 December 2024  10:09:09 -0500 (0:00:00.046)       0:01:43.778 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Wednesday 11 December 2024  10:09:09 -0500 (0:00:00.085)       0:01:43.864 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Wednesday 11 December 2024  10:09:09 -0500 (0:00:00.043)       0:01:43.908 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sda",
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Wednesday 11 December 2024  10:09:10 -0500 (0:00:00.481)       0:01:44.389 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Wednesday 11 December 2024  10:09:10 -0500 (0:00:00.040)       0:01:44.429 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Wednesday 11 December 2024  10:09:10 -0500 (0:00:00.046)       0:01:44.476 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Wednesday 11 December 2024  10:09:10 -0500 (0:00:00.035)       0:01:44.512 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Wednesday 11 December 2024  10:09:10 -0500 (0:00:00.039)       0:01:44.551 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Wednesday 11 December 2024  10:09:10 -0500 (0:00:00.040)       0:01:44.592 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Wednesday 11 December 2024  10:09:10 -0500 (0:00:00.034)       0:01:44.626 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Wednesday 11 December 2024  10:09:10 -0500 (0:00:00.055)       0:01:44.682 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.41.57 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Wednesday 11 December 2024  10:09:10 -0500 (0:00:00.489)       0:01:45.172 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Wednesday 11 December 2024  10:09:10 -0500 (0:00:00.041)       0:01:45.213 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node3

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.087)       0:01:45.301 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.037)       0:01:45.338 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.033)       0:01:45.372 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.033)       0:01:45.405 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.034)       0:01:45.440 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.035)       0:01:45.475 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.037)       0:01:45.513 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.033)       0:01:45.547 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.031)       0:01:45.579 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.025)       0:01:45.604 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.024)       0:01:45.628 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.025)       0:01:45.654 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node3

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.056)       0:01:45.710 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node3

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.051)       0:01:45.762 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.055)       0:01:45.818 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.022)       0:01:45.841 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.022)       0:01:45.863 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.023)       0:01:45.887 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.024)       0:01:45.911 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.021)       0:01:45.933 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.023)       0:01:45.957 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node3

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.044)       0:01:46.001 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node3

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.043)       0:01:46.045 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.021)       0:01:46.066 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.021)       0:01:46.088 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.021)       0:01:46.109 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.030)       0:01:46.139 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.061)       0:01:46.201 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.045)       0:01:46.246 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Wednesday 11 December 2024  10:09:11 -0500 (0:00:00.035)       0:01:46.282 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node3

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.064)       0:01:46.347 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.037)       0:01:46.384 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.029)       0:01:46.413 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.025)       0:01:46.439 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.026)       0:01:46.466 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.023)       0:01:46.489 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.027)       0:01:46.517 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.022)       0:01:46.539 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node3

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.050)       0:01:46.590 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node3

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.044)       0:01:46.635 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.023)       0:01:46.658 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.021)       0:01:46.680 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.021)       0:01:46.701 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.021)       0:01:46.722 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.042)       0:01:46.765 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.028)       0:01:46.794 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.022)       0:01:46.816 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node3

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.068)       0:01:46.885 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.024)       0:01:46.910 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.028)       0:01:46.938 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.022)       0:01:46.961 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.028)       0:01:46.989 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.022)       0:01:47.011 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.023)       0:01:47.035 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.021)       0:01:47.056 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node3

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.040)       0:01:47.097 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Wednesday 11 December 2024  10:09:12 -0500 (0:00:00.026)       0:01:47.123 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node3

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.180)       0:01:47.304 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.028)       0:01:47.333 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.032)       0:01:47.365 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.028)       0:01:47.394 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.028)       0:01:47.423 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.024)       0:01:47.447 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.023)       0:01:47.470 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.020)       0:01:47.491 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.021)       0:01:47.513 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.020)       0:01:47.533 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.020)       0:01:47.553 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.019)       0:01:47.573 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.052)       0:01:47.625 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.037)       0:01:47.663 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.038)       0:01:47.701 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.034)       0:01:47.736 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.046)       0:01:47.783 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.037)       0:01:47.820 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.044)       0:01:47.864 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Wednesday 11 December 2024  10:09:13 -0500 (0:00:00.047)       0:01:47.912 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929742.9714515,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1733929742.9714515,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 287887,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1733929742.9714515,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Wednesday 11 December 2024  10:09:14 -0500 (0:00:00.449)       0:01:48.361 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Wednesday 11 December 2024  10:09:14 -0500 (0:00:00.038)       0:01:48.399 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Wednesday 11 December 2024  10:09:14 -0500 (0:00:00.023)       0:01:48.423 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Wednesday 11 December 2024  10:09:14 -0500 (0:00:00.039)       0:01:48.462 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Wednesday 11 December 2024  10:09:14 -0500 (0:00:00.041)       0:01:48.504 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Wednesday 11 December 2024  10:09:14 -0500 (0:00:00.031)       0:01:48.535 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Wednesday 11 December 2024  10:09:14 -0500 (0:00:00.041)       0:01:48.577 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Wednesday 11 December 2024  10:09:14 -0500 (0:00:00.034)       0:01:48.612 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Wednesday 11 December 2024  10:09:17 -0500 (0:00:03.015)       0:01:51.628 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Wednesday 11 December 2024  10:09:17 -0500 (0:00:00.046)       0:01:51.674 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Wednesday 11 December 2024  10:09:17 -0500 (0:00:00.056)       0:01:51.731 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Wednesday 11 December 2024  10:09:17 -0500 (0:00:00.068)       0:01:51.799 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Wednesday 11 December 2024  10:09:17 -0500 (0:00:00.057)       0:01:51.857 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Wednesday 11 December 2024  10:09:17 -0500 (0:00:00.035)       0:01:51.892 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Wednesday 11 December 2024  10:09:17 -0500 (0:00:00.064)       0:01:51.957 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Wednesday 11 December 2024  10:09:17 -0500 (0:00:00.054)       0:01:52.011 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Wednesday 11 December 2024  10:09:17 -0500 (0:00:00.073)       0:01:52.084 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Wednesday 11 December 2024  10:09:17 -0500 (0:00:00.076)       0:01:52.161 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Wednesday 11 December 2024  10:09:17 -0500 (0:00:00.042)       0:01:52.203 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Wednesday 11 December 2024  10:09:17 -0500 (0:00:00.074)       0:01:52.278 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Wednesday 11 December 2024  10:09:18 -0500 (0:00:00.037)       0:01:52.316 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Wednesday 11 December 2024  10:09:18 -0500 (0:00:00.044)       0:01:52.361 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Wednesday 11 December 2024  10:09:18 -0500 (0:00:00.047)       0:01:52.409 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Wednesday 11 December 2024  10:09:18 -0500 (0:00:00.034)       0:01:52.443 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Wednesday 11 December 2024  10:09:18 -0500 (0:00:00.034)       0:01:52.478 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Wednesday 11 December 2024  10:09:18 -0500 (0:00:00.034)       0:01:52.512 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Wednesday 11 December 2024  10:09:18 -0500 (0:00:00.036)       0:01:52.548 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Wednesday 11 December 2024  10:09:18 -0500 (0:00:00.032)       0:01:52.581 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Wednesday 11 December 2024  10:09:18 -0500 (0:00:00.074)       0:01:52.655 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Wednesday 11 December 2024  10:09:18 -0500 (0:00:00.098)       0:01:52.754 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Wednesday 11 December 2024  10:09:18 -0500 (0:00:00.078)       0:01:52.832 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Wednesday 11 December 2024  10:09:18 -0500 (0:00:00.033)       0:01:52.866 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Wednesday 11 December 2024  10:09:18 -0500 (0:00:00.037)       0:01:52.903 **** 
ok: [managed-node3] => {
    "bytes": 10737418240,
    "changed": false,
    "lvm": "10g",
    "parted": "10GiB",
    "size": "10 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Wednesday 11 December 2024  10:09:19 -0500 (0:00:00.451)       0:01:53.355 **** 
ok: [managed-node3] => {
    "bytes": 10737418240,
    "changed": false,
    "lvm": "10g",
    "parted": "10GiB",
    "size": "10 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Wednesday 11 December 2024  10:09:19 -0500 (0:00:00.413)       0:01:53.768 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_expected_size": "10737418240"
    },
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Wednesday 11 December 2024  10:09:19 -0500 (0:00:00.069)       0:01:53.838 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "10737418240"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Wednesday 11 December 2024  10:09:19 -0500 (0:00:00.166)       0:01:54.004 **** 
ok: [managed-node3] => {
    "bytes": 10726680821,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.500)       0:01:54.516 **** 
skipping: [managed-node3] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.026)       0:01:54.542 **** 
skipping: [managed-node3] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.028)       0:01:54.571 **** 
skipping: [managed-node3] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.037)       0:01:54.608 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.039)       0:01:54.647 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.031)       0:01:54.679 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.036)       0:01:54.716 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.028)       0:01:54.744 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.034)       0:01:54.779 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.032)       0:01:54.811 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.031)       0:01:54.842 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.032)       0:01:54.875 **** 
skipping: [managed-node3] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.034)       0:01:54.909 **** 
skipping: [managed-node3] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.031)       0:01:54.940 **** 
skipping: [managed-node3] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.030)       0:01:54.971 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.023)       0:01:54.994 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.024)       0:01:55.019 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.022)       0:01:55.041 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.039)       0:01:55.080 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.035)       0:01:55.115 **** 
ok: [managed-node3] => {
    "storage_test_actual_size": {
        "bytes": 10737418240,
        "changed": false,
        "failed": false,
        "lvm": "10g",
        "parted": "10GiB",
        "size": "10 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.040)       0:01:55.156 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "10737418240"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.034)       0:01:55.190 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Wednesday 11 December 2024  10:09:20 -0500 (0:00:00.039)       0:01:55.230 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.022192",
    "end": "2024-12-11 10:09:21.284830",
    "rc": 0,
    "start": "2024-12-11 10:09:21.262638"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.410)       0:01:55.641 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.028)       0:01:55.669 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.031)       0:01:55.701 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.025)       0:01:55.726 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.024)       0:01:55.750 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.024)       0:01:55.775 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.023)       0:01:55.798 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.021)       0:01:55.819 **** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.021)       0:01:55.841 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Test for correct handling of invalid size specification] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:128
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.020)       0:01:55.862 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node3

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.034)       0:01:55.897 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_pools_global": [],
        "storage_safe_mode_global": false,
        "storage_volumes_global": []
    },
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.027)       0:01:55.924 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.030)       0:01:55.955 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.031)       0:01:55.987 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.026)       0:01:56.013 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.060)       0:01:56.074 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.022)       0:01:56.096 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.021)       0:01:56.118 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.019)       0:01:56.137 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.019)       0:01:56.157 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.051)       0:01:56.208 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.022)       0:01:56.231 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "xyz GiB"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.027)       0:01:56.259 **** 
ok: [managed-node3] => {
    "storage_volumes": []
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:09:21 -0500 (0:00:00.027)       0:01:56.286 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:09:22 -0500 (0:00:00.058)       0:01:56.345 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:09:22 -0500 (0:00:00.023)       0:01:56.368 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:09:22 -0500 (0:00:00.023)       0:01:56.391 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:09:22 -0500 (0:00:00.022)       0:01:56.414 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:09:22 -0500 (0:00:00.036)       0:01:56.450 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:09:22 -0500 (0:00:00.027)       0:01:56.478 **** 
fatal: [managed-node3]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

invalid size specification 'xyz GiB' in pool 'foo'

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Wednesday 11 December 2024  10:09:26 -0500 (0:00:04.221)       0:02:00.699 **** 
fatal: [managed-node3]: FAILED! => {
    "changed": false
}

MSG:

{'msg': "invalid size specification 'xyz GiB' in pool 'foo'", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'encryption_clevis_pin': None, 'encryption_tang_url': None, 'encryption_tang_thumbprint': None, 'grow_to_fill': False, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'shared': False, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': None, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': None, 'fs_label': None, 'fs_type': 'ext4', 'mount_options': None, 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': 'xyz GiB', 'state': 'present', 'type': None, 'cached': None, 'cache_devices': [], 'cache_mode': None, 'cache_size': None, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None}]}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'grow_to_fill': False, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, 'shared': False}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': False, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.029)       0:02:00.729 **** 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.020)       0:02:00.750 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.024)       0:02:00.775 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.032)       0:02:00.808 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Test for correct handling of invalid size specification] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:144
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.022)       0:02:00.830 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node3

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.035)       0:02:00.865 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_pools_global": [],
        "storage_safe_mode_global": false,
        "storage_volumes_global": []
    },
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.027)       0:02:00.893 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.032)       0:02:00.925 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.032)       0:02:00.957 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.026)       0:02:00.984 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.060)       0:02:01.044 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.022)       0:02:01.067 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.021)       0:02:01.089 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.021)       0:02:01.110 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.020)       0:02:01.130 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.051)       0:02:01.182 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.023)       0:02:01.206 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "none"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.029)       0:02:01.235 **** 
ok: [managed-node3] => {
    "storage_volumes": []
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.028)       0:02:01.264 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:09:26 -0500 (0:00:00.024)       0:02:01.289 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:09:27 -0500 (0:00:00.022)       0:02:01.311 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:09:27 -0500 (0:00:00.023)       0:02:01.335 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:09:27 -0500 (0:00:00.023)       0:02:01.358 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:09:27 -0500 (0:00:00.033)       0:02:01.392 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:09:27 -0500 (0:00:00.020)       0:02:01.412 **** 
fatal: [managed-node3]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

invalid size specification 'none' in pool 'foo'

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Wednesday 11 December 2024  10:09:31 -0500 (0:00:04.123)       0:02:05.536 **** 
fatal: [managed-node3]: FAILED! => {
    "changed": false
}

MSG:

{'msg': "invalid size specification 'none' in pool 'foo'", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'encryption_clevis_pin': None, 'encryption_tang_url': None, 'encryption_tang_thumbprint': None, 'grow_to_fill': False, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'shared': False, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': None, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': None, 'fs_label': None, 'fs_type': 'ext4', 'mount_options': None, 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': 'none', 'state': 'present', 'type': None, 'cached': None, 'cache_devices': [], 'cache_mode': None, 'cache_size': None, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None}]}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'grow_to_fill': False, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, 'shared': False}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': False, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.036)       0:02:05.573 **** 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.028)       0:02:05.601 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.039)       0:02:05.640 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.045)       0:02:05.685 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up] ****************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:160
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.024)       0:02:05.710 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.066)       0:02:05.777 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.035)       0:02:05.812 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.035)       0:02:05.847 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.063)       0:02:05.911 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.060)       0:02:05.971 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.022)       0:02:05.994 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.021)       0:02:06.015 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.021)       0:02:06.036 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.050)       0:02:06.087 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.022)       0:02:06.110 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "state": "absent",
            "volumes": [
                {
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "5g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.027)       0:02:06.137 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.022)       0:02:06.160 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.022)       0:02:06.182 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.023)       0:02:06.206 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.022)       0:02:06.228 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.021)       0:02:06.250 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:09:31 -0500 (0:00:00.034)       0:02:06.285 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:09:32 -0500 (0:00:00.020)       0:02:06.305 **** 
changed: [managed-node3] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sda",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "absent",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:09:36 -0500 (0:00:04.573)       0:02:10.878 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:09:36 -0500 (0:00:00.030)       0:02:10.909 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929671.0876713,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "306e713ef45e6db45ca0244c07767212954e6abb",
        "ctime": 1733929671.0846713,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929671.0846713,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1394,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:09:36 -0500 (0:00:00.385)       0:02:11.295 **** 
ok: [managed-node3] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:09:37 -0500 (0:00:00.377)       0:02:11.672 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:09:37 -0500 (0:00:00.021)       0:02:11.693 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sda",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "absent",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:09:37 -0500 (0:00:00.027)       0:02:11.721 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "absent",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:09:37 -0500 (0:00:00.025)       0:02:11.747 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:09:37 -0500 (0:00:00.025)       0:02:11.772 **** 
changed: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:09:37 -0500 (0:00:00.395)       0:02:12.168 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:09:38 -0500 (0:00:00.634)       0:02:12.802 **** 

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:09:38 -0500 (0:00:00.022)       0:02:12.824 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:09:38 -0500 (0:00:00.021)       0:02:12.846 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:09:39 -0500 (0:00:00.648)       0:02:13.495 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:09:39 -0500 (0:00:00.383)       0:02:13.878 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:09:39 -0500 (0:00:00.020)       0:02:13.898 **** 
ok: [managed-node3]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:173
Wednesday 11 December 2024  10:09:40 -0500 (0:00:00.762)       0:02:14.661 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node3

TASK [Print out pool information] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Wednesday 11 December 2024  10:09:40 -0500 (0:00:00.074)       0:02:14.736 **** 
ok: [managed-node3] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "absent",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Wednesday 11 December 2024  10:09:40 -0500 (0:00:00.029)       0:02:14.765 **** 
skipping: [managed-node3] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Wednesday 11 December 2024  10:09:40 -0500 (0:00:00.022)       0:02:14.788 **** 
ok: [managed-node3] => {
    "changed": false,
    "info": {
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Wednesday 11 December 2024  10:09:40 -0500 (0:00:00.370)       0:02:15.158 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002858",
    "end": "2024-12-11 10:09:41.173234",
    "rc": 0,
    "start": "2024-12-11 10:09:41.170376"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Wednesday 11 December 2024  10:09:41 -0500 (0:00:00.369)       0:02:15.528 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002470",
    "end": "2024-12-11 10:09:41.539280",
    "failed_when_result": false,
    "rc": 0,
    "start": "2024-12-11 10:09:41.536810"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Wednesday 11 December 2024  10:09:41 -0500 (0:00:00.365)       0:02:15.893 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node3

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Wednesday 11 December 2024  10:09:41 -0500 (0:00:00.042)       0:02:15.936 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Wednesday 11 December 2024  10:09:41 -0500 (0:00:00.019)       0:02:15.956 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Wednesday 11 December 2024  10:09:41 -0500 (0:00:00.020)       0:02:15.976 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Wednesday 11 December 2024  10:09:41 -0500 (0:00:00.020)       0:02:15.997 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Wednesday 11 December 2024  10:09:41 -0500 (0:00:00.060)       0:02:16.057 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "0",
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Wednesday 11 December 2024  10:09:41 -0500 (0:00:00.030)       0:02:16.088 **** 

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Wednesday 11 December 2024  10:09:41 -0500 (0:00:00.019)       0:02:16.108 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": "0"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Wednesday 11 December 2024  10:09:41 -0500 (0:00:00.027)       0:02:16.135 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": []
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Wednesday 11 December 2024  10:09:41 -0500 (0:00:00.026)       0:02:16.161 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Wednesday 11 December 2024  10:09:41 -0500 (0:00:00.026)       0:02:16.188 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Wednesday 11 December 2024  10:09:41 -0500 (0:00:00.026)       0:02:16.215 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Wednesday 11 December 2024  10:09:41 -0500 (0:00:00.025)       0:02:16.240 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Wednesday 11 December 2024  10:09:41 -0500 (0:00:00.021)       0:02:16.261 **** 

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Wednesday 11 December 2024  10:09:41 -0500 (0:00:00.018)       0:02:16.280 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.41.57 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.408)       0:02:16.689 **** 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.021)       0:02:16.710 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node3

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.041)       0:02:16.751 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.022)       0:02:16.774 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.021)       0:02:16.796 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.021)       0:02:16.818 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.023)       0:02:16.842 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.021)       0:02:16.863 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.020)       0:02:16.883 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.021)       0:02:16.905 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.019)       0:02:16.924 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.019)       0:02:16.944 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.022)       0:02:16.967 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.019)       0:02:16.986 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node3

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.041)       0:02:17.028 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node3

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.051)       0:02:17.079 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.056)       0:02:17.136 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.021)       0:02:17.157 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.021)       0:02:17.179 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.021)       0:02:17.201 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.021)       0:02:17.222 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.020)       0:02:17.242 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Wednesday 11 December 2024  10:09:42 -0500 (0:00:00.023)       0:02:17.265 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node3

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.042)       0:02:17.308 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node3

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.043)       0:02:17.351 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.022)       0:02:17.373 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.021)       0:02:17.395 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.021)       0:02:17.416 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.020)       0:02:17.436 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.047)       0:02:17.484 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.025)       0:02:17.509 **** 

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.018)       0:02:17.528 **** 

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.018)       0:02:17.547 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.020)       0:02:17.567 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node3

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.047)       0:02:17.615 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node3

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.043)       0:02:17.659 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.022)       0:02:17.681 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.022)       0:02:17.703 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.021)       0:02:17.725 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.021)       0:02:17.747 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.021)       0:02:17.768 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.021)       0:02:17.790 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.022)       0:02:17.813 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node3

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.051)       0:02:17.864 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.021)       0:02:17.885 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.020)       0:02:17.906 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.021)       0:02:17.928 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.021)       0:02:17.949 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.021)       0:02:17.971 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.019)       0:02:17.990 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.020)       0:02:18.011 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node3

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.039)       0:02:18.050 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": false,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.027)       0:02:18.078 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node3

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.122)       0:02:18.200 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.026)       0:02:18.227 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.028)       0:02:18.255 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.023)       0:02:18.278 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Wednesday 11 December 2024  10:09:43 -0500 (0:00:00.017)       0:02:18.295 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.021)       0:02:18.317 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.021)       0:02:18.338 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.021)       0:02:18.360 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.021)       0:02:18.381 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.063)       0:02:18.445 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.022)       0:02:18.467 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.020)       0:02:18.488 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "0",
        "storage_test_fstab_expected_mount_options_matches": "0",
        "storage_test_fstab_expected_mount_point_matches": "0",
        "storage_test_fstab_id_matches": [],
        "storage_test_fstab_mount_options_matches": [],
        "storage_test_fstab_mount_point_matches": []
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.044)       0:02:18.533 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.022)       0:02:18.555 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.026)       0:02:18.581 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.023)       0:02:18.605 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.027)       0:02:18.633 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.022)       0:02:18.655 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.024)       0:02:18.680 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.024)       0:02:18.704 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "exists": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.364)       0:02:19.068 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.022)       0:02:19.091 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.027)       0:02:19.118 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.023)       0:02:19.142 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.027)       0:02:19.170 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.020)       0:02:19.191 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.016)       0:02:19.208 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Wednesday 11 December 2024  10:09:44 -0500 (0:00:00.022)       0:02:19.230 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Wednesday 11 December 2024  10:09:47 -0500 (0:00:02.833)       0:02:22.064 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Wednesday 11 December 2024  10:09:47 -0500 (0:00:00.024)       0:02:22.089 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Wednesday 11 December 2024  10:09:47 -0500 (0:00:00.025)       0:02:22.115 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Wednesday 11 December 2024  10:09:47 -0500 (0:00:00.023)       0:02:22.138 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Wednesday 11 December 2024  10:09:47 -0500 (0:00:00.025)       0:02:22.163 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Wednesday 11 December 2024  10:09:47 -0500 (0:00:00.020)       0:02:22.183 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Wednesday 11 December 2024  10:09:47 -0500 (0:00:00.026)       0:02:22.209 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Wednesday 11 December 2024  10:09:47 -0500 (0:00:00.019)       0:02:22.229 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Wednesday 11 December 2024  10:09:47 -0500 (0:00:00.016)       0:02:22.246 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Wednesday 11 December 2024  10:09:47 -0500 (0:00:00.030)       0:02:22.276 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.025)       0:02:22.302 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.021)       0:02:22.323 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.020)       0:02:22.343 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.024)       0:02:22.368 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.019)       0:02:22.387 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.031)       0:02:22.419 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.035)       0:02:22.454 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.035)       0:02:22.490 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.033)       0:02:22.523 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.031)       0:02:22.555 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.034)       0:02:22.589 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.035)       0:02:22.624 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.032)       0:02:22.657 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.033)       0:02:22.691 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.033)       0:02:22.724 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.025)       0:02:22.750 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.024)       0:02:22.774 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.025)       0:02:22.799 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "10737418240"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.027)       0:02:22.827 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.025)       0:02:22.852 **** 
skipping: [managed-node3] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.024)       0:02:22.876 **** 
skipping: [managed-node3] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.024)       0:02:22.900 **** 
skipping: [managed-node3] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.022)       0:02:22.923 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.024)       0:02:22.947 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.020)       0:02:22.968 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.020)       0:02:22.989 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.022)       0:02:23.011 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.022)       0:02:23.034 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.021)       0:02:23.055 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.022)       0:02:23.077 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.063)       0:02:23.141 **** 
skipping: [managed-node3] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.034)       0:02:23.175 **** 
skipping: [managed-node3] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.031)       0:02:23.207 **** 
skipping: [managed-node3] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.029)       0:02:23.236 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.032)       0:02:23.268 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Wednesday 11 December 2024  10:09:48 -0500 (0:00:00.028)       0:02:23.296 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.027)       0:02:23.324 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.028)       0:02:23.352 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.022)       0:02:23.374 **** 
ok: [managed-node3] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.026)       0:02:23.401 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "10737418240"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.023)       0:02:23.425 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.022)       0:02:23.447 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.023)       0:02:23.471 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.022)       0:02:23.493 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.022)       0:02:23.515 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.022)       0:02:23.538 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.020)       0:02:23.558 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.022)       0:02:23.581 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.023)       0:02:23.605 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.021)       0:02:23.626 **** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.019)       0:02:23.646 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Create a LVM logical volume with for ext3 FS size 5g] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:178
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.022)       0:02:23.668 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.053)       0:02:23.722 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.029)       0:02:23.752 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.025)       0:02:23.777 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.059)       0:02:23.837 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.022)       0:02:23.859 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.021)       0:02:23.881 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.020)       0:02:23.902 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.028)       0:02:23.930 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.071)       0:02:24.002 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.044)       0:02:24.046 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext3",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "5g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.040)       0:02:24.087 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.037)       0:02:24.124 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.033)       0:02:24.158 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.025)       0:02:24.183 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.027)       0:02:24.211 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:09:49 -0500 (0:00:00.028)       0:02:24.240 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:09:50 -0500 (0:00:00.059)       0:02:24.299 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:09:50 -0500 (0:00:00.023)       0:02:24.322 **** 
changed: [managed-node3] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/sda",
            "fs_type": "lvmpv"
        },
        {
            "action": "create device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "create device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext3"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1",
        "/dev/mapper/foo-test1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext3",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs",
        "lvm2"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext3",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:09:56 -0500 (0:00:06.711)       0:02:31.033 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:09:56 -0500 (0:00:00.021)       0:02:31.055 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929778.332281,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "ab8070345774adad92683e9645714452be7be474",
        "ctime": 1733929777.8082838,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929777.8082838,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1343,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:09:57 -0500 (0:00:00.396)       0:02:31.451 **** 
ok: [managed-node3] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:09:57 -0500 (0:00:00.421)       0:02:31.873 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:09:57 -0500 (0:00:00.021)       0:02:31.895 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/sda",
                "fs_type": "lvmpv"
            },
            {
                "action": "create device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "create device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext3"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1",
            "/dev/mapper/foo-test1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext3",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs",
            "lvm2"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext3",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:09:57 -0500 (0:00:00.027)       0:02:31.922 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext3",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:09:57 -0500 (0:00:00.027)       0:02:31.950 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:09:57 -0500 (0:00:00.025)       0:02:31.976 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:09:57 -0500 (0:00:00.022)       0:02:31.998 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:09:58 -0500 (0:00:00.636)       0:02:32.635 **** 
changed: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext3', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext3",
    "mount_info": {
        "dump": 0,
        "fstype": "ext3",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:09:58 -0500 (0:00:00.455)       0:02:33.090 **** 
skipping: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext3', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "ext3",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:09:58 -0500 (0:00:00.030)       0:02:33.120 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:09:59 -0500 (0:00:00.644)       0:02:33.765 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:09:59 -0500 (0:00:00.384)       0:02:34.150 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:09:59 -0500 (0:00:00.029)       0:02:34.180 **** 
ok: [managed-node3]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:194
Wednesday 11 December 2024  10:10:00 -0500 (0:00:00.793)       0:02:34.973 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node3

TASK [Print out pool information] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Wednesday 11 December 2024  10:10:00 -0500 (0:00:00.046)       0:02:35.019 **** 
ok: [managed-node3] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext3",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Wednesday 11 December 2024  10:10:00 -0500 (0:00:00.028)       0:02:35.048 **** 
skipping: [managed-node3] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Wednesday 11 December 2024  10:10:00 -0500 (0:00:00.022)       0:02:35.070 **** 
ok: [managed-node3] => {
    "changed": false,
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "ext3",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/foo-test1",
            "size": "5G",
            "type": "lvm",
            "uuid": "c7a0b1fc-e4b0-46bc-b21a-e0e251d60a13"
        },
        "/dev/sda": {
            "fstype": "LVM2_member",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "dfVLB2-vpGU-RqJQ-8Pgk-QQh5-6F0Q-KfgQjh"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Wednesday 11 December 2024  10:10:01 -0500 (0:00:00.388)       0:02:35.459 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003224",
    "end": "2024-12-11 10:10:01.498637",
    "rc": 0,
    "start": "2024-12-11 10:10:01.495413"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/foo-test1 /opt/test1 ext3 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Wednesday 11 December 2024  10:10:01 -0500 (0:00:00.408)       0:02:35.867 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003183",
    "end": "2024-12-11 10:10:01.904158",
    "failed_when_result": false,
    "rc": 0,
    "start": "2024-12-11 10:10:01.900975"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Wednesday 11 December 2024  10:10:01 -0500 (0:00:00.422)       0:02:36.289 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node3

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Wednesday 11 December 2024  10:10:02 -0500 (0:00:00.066)       0:02:36.355 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Wednesday 11 December 2024  10:10:02 -0500 (0:00:00.026)       0:02:36.382 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "vgs",
        "--noheadings",
        "--binary",
        "-o",
        "shared",
        "foo"
    ],
    "delta": "0:00:00.025234",
    "end": "2024-12-11 10:10:02.474577",
    "rc": 0,
    "start": "2024-12-11 10:10:02.449343"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Wednesday 11 December 2024  10:10:02 -0500 (0:00:00.448)       0:02:36.831 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Wednesday 11 December 2024  10:10:02 -0500 (0:00:00.032)       0:02:36.863 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Wednesday 11 December 2024  10:10:02 -0500 (0:00:00.047)       0:02:36.911 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Wednesday 11 December 2024  10:10:02 -0500 (0:00:00.035)       0:02:36.946 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sda",
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Wednesday 11 December 2024  10:10:03 -0500 (0:00:00.433)       0:02:37.380 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Wednesday 11 December 2024  10:10:03 -0500 (0:00:00.032)       0:02:37.413 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Wednesday 11 December 2024  10:10:03 -0500 (0:00:00.069)       0:02:37.482 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Wednesday 11 December 2024  10:10:03 -0500 (0:00:00.037)       0:02:37.520 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Wednesday 11 December 2024  10:10:03 -0500 (0:00:00.031)       0:02:37.551 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Wednesday 11 December 2024  10:10:03 -0500 (0:00:00.036)       0:02:37.588 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Wednesday 11 December 2024  10:10:03 -0500 (0:00:00.026)       0:02:37.614 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Wednesday 11 December 2024  10:10:03 -0500 (0:00:00.050)       0:02:37.665 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.41.57 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Wednesday 11 December 2024  10:10:03 -0500 (0:00:00.479)       0:02:38.145 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Wednesday 11 December 2024  10:10:03 -0500 (0:00:00.056)       0:02:38.201 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node3

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Wednesday 11 December 2024  10:10:03 -0500 (0:00:00.074)       0:02:38.275 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.033)       0:02:38.308 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.033)       0:02:38.342 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.045)       0:02:38.388 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.042)       0:02:38.430 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.040)       0:02:38.471 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.037)       0:02:38.508 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.040)       0:02:38.549 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.029)       0:02:38.578 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.035)       0:02:38.613 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.031)       0:02:38.645 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.034)       0:02:38.679 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node3

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.092)       0:02:38.771 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node3

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.081)       0:02:38.853 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.052)       0:02:38.905 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.034)       0:02:38.940 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.032)       0:02:38.972 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.033)       0:02:39.006 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.067)       0:02:39.073 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.036)       0:02:39.110 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.037)       0:02:39.147 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node3

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Wednesday 11 December 2024  10:10:04 -0500 (0:00:00.080)       0:02:39.228 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node3

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.077)       0:02:39.305 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.034)       0:02:39.339 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.033)       0:02:39.373 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.045)       0:02:39.418 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.044)       0:02:39.462 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.075)       0:02:39.538 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.036)       0:02:39.574 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.039)       0:02:39.614 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node3

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.065)       0:02:39.679 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.034)       0:02:39.714 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.054)       0:02:39.768 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.048)       0:02:39.817 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.034)       0:02:39.851 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.034)       0:02:39.885 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.038)       0:02:39.924 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.033)       0:02:39.957 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node3

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.081)       0:02:40.039 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node3

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.121)       0:02:40.160 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.027)       0:02:40.187 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.029)       0:02:40.217 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.032)       0:02:40.249 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Wednesday 11 December 2024  10:10:05 -0500 (0:00:00.034)       0:02:40.283 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.038)       0:02:40.322 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.041)       0:02:40.363 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.035)       0:02:40.398 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node3

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.097)       0:02:40.495 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.034)       0:02:40.529 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.032)       0:02:40.562 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.034)       0:02:40.596 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.035)       0:02:40.631 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.034)       0:02:40.666 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.035)       0:02:40.702 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.033)       0:02:40.736 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node3

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.060)       0:02:40.797 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.057)       0:02:40.855 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node3

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.170)       0:02:41.025 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.039)       0:02:41.065 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.042)       0:02:41.108 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.032)       0:02:41.140 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.041)       0:02:41.182 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.036)       0:02:41.218 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.032)       0:02:41.251 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Wednesday 11 December 2024  10:10:06 -0500 (0:00:00.033)       0:02:41.284 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Wednesday 11 December 2024  10:10:07 -0500 (0:00:00.032)       0:02:41.316 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Wednesday 11 December 2024  10:10:07 -0500 (0:00:00.034)       0:02:41.351 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Wednesday 11 December 2024  10:10:07 -0500 (0:00:00.033)       0:02:41.385 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Wednesday 11 December 2024  10:10:07 -0500 (0:00:00.036)       0:02:41.422 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext3 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Wednesday 11 December 2024  10:10:07 -0500 (0:00:00.070)       0:02:41.492 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Wednesday 11 December 2024  10:10:07 -0500 (0:00:00.038)       0:02:41.530 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Wednesday 11 December 2024  10:10:07 -0500 (0:00:00.041)       0:02:41.571 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Wednesday 11 December 2024  10:10:07 -0500 (0:00:00.034)       0:02:41.605 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Wednesday 11 December 2024  10:10:07 -0500 (0:00:00.039)       0:02:41.645 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Wednesday 11 December 2024  10:10:07 -0500 (0:00:00.036)       0:02:41.681 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Wednesday 11 December 2024  10:10:07 -0500 (0:00:00.046)       0:02:41.728 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Wednesday 11 December 2024  10:10:07 -0500 (0:00:00.049)       0:02:41.777 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929796.622193,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1733929796.622193,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 321740,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1733929796.622193,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Wednesday 11 December 2024  10:10:07 -0500 (0:00:00.449)       0:02:42.226 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Wednesday 11 December 2024  10:10:07 -0500 (0:00:00.042)       0:02:42.269 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Wednesday 11 December 2024  10:10:08 -0500 (0:00:00.036)       0:02:42.306 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Wednesday 11 December 2024  10:10:08 -0500 (0:00:00.046)       0:02:42.353 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Wednesday 11 December 2024  10:10:08 -0500 (0:00:00.038)       0:02:42.391 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Wednesday 11 December 2024  10:10:08 -0500 (0:00:00.037)       0:02:42.429 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Wednesday 11 December 2024  10:10:08 -0500 (0:00:00.040)       0:02:42.470 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Wednesday 11 December 2024  10:10:08 -0500 (0:00:00.032)       0:02:42.502 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Wednesday 11 December 2024  10:10:11 -0500 (0:00:02.882)       0:02:45.384 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.027)       0:02:45.412 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.031)       0:02:45.444 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.040)       0:02:45.485 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.035)       0:02:45.520 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.032)       0:02:45.553 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.041)       0:02:45.594 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.034)       0:02:45.629 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.058)       0:02:45.688 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.046)       0:02:45.735 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.064)       0:02:45.800 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.068)       0:02:45.868 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.035)       0:02:45.903 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.046)       0:02:45.950 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.044)       0:02:45.995 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.044)       0:02:46.039 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.032)       0:02:46.072 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.039)       0:02:46.112 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.033)       0:02:46.145 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.054)       0:02:46.199 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Wednesday 11 December 2024  10:10:11 -0500 (0:00:00.036)       0:02:46.236 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Wednesday 11 December 2024  10:10:12 -0500 (0:00:00.068)       0:02:46.305 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Wednesday 11 December 2024  10:10:12 -0500 (0:00:00.036)       0:02:46.341 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Wednesday 11 December 2024  10:10:12 -0500 (0:00:00.030)       0:02:46.372 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Wednesday 11 December 2024  10:10:12 -0500 (0:00:00.035)       0:02:46.407 **** 
ok: [managed-node3] => {
    "bytes": 5368709120,
    "changed": false,
    "lvm": "5g",
    "parted": "5GiB",
    "size": "5 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Wednesday 11 December 2024  10:10:12 -0500 (0:00:00.515)       0:02:46.922 **** 
ok: [managed-node3] => {
    "bytes": 5368709120,
    "changed": false,
    "lvm": "5g",
    "parted": "5GiB",
    "size": "5 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Wednesday 11 December 2024  10:10:13 -0500 (0:00:00.652)       0:02:47.575 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_expected_size": "5368709120"
    },
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Wednesday 11 December 2024  10:10:13 -0500 (0:00:00.041)       0:02:47.616 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Wednesday 11 December 2024  10:10:13 -0500 (0:00:00.035)       0:02:47.652 **** 
ok: [managed-node3] => {
    "bytes": 10726680821,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Wednesday 11 December 2024  10:10:13 -0500 (0:00:00.410)       0:02:48.062 **** 
skipping: [managed-node3] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Wednesday 11 December 2024  10:10:13 -0500 (0:00:00.031)       0:02:48.093 **** 
skipping: [managed-node3] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Wednesday 11 December 2024  10:10:13 -0500 (0:00:00.025)       0:02:48.119 **** 
skipping: [managed-node3] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Wednesday 11 December 2024  10:10:13 -0500 (0:00:00.025)       0:02:48.144 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Wednesday 11 December 2024  10:10:13 -0500 (0:00:00.026)       0:02:48.171 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Wednesday 11 December 2024  10:10:13 -0500 (0:00:00.022)       0:02:48.193 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Wednesday 11 December 2024  10:10:13 -0500 (0:00:00.023)       0:02:48.216 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Wednesday 11 December 2024  10:10:13 -0500 (0:00:00.022)       0:02:48.238 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Wednesday 11 December 2024  10:10:13 -0500 (0:00:00.022)       0:02:48.260 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Wednesday 11 December 2024  10:10:13 -0500 (0:00:00.020)       0:02:48.281 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.024)       0:02:48.305 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.023)       0:02:48.329 **** 
skipping: [managed-node3] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.021)       0:02:48.351 **** 
skipping: [managed-node3] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.021)       0:02:48.372 **** 
skipping: [managed-node3] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.021)       0:02:48.393 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.021)       0:02:48.415 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.024)       0:02:48.439 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.022)       0:02:48.461 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.022)       0:02:48.484 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.022)       0:02:48.506 **** 
ok: [managed-node3] => {
    "storage_test_actual_size": {
        "bytes": 5368709120,
        "changed": false,
        "failed": false,
        "lvm": "5g",
        "parted": "5GiB",
        "size": "5 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.024)       0:02:48.531 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.023)       0:02:48.554 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.037)       0:02:48.592 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.024016",
    "end": "2024-12-11 10:10:14.639997",
    "rc": 0,
    "start": "2024-12-11 10:10:14.615981"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.405)       0:02:48.997 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.028)       0:02:49.025 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.029)       0:02:49.055 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.025)       0:02:49.081 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.034)       0:02:49.115 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.026)       0:02:49.141 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.025)       0:02:49.166 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.021)       0:02:49.188 **** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.019)       0:02:49.207 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Change volume size to 9g] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:197
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.022)       0:02:49.230 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:10:14 -0500 (0:00:00.057)       0:02:49.288 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.032)       0:02:49.320 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.104)       0:02:49.425 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.057)       0:02:49.483 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.023)       0:02:49.506 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.021)       0:02:49.527 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.021)       0:02:49.549 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.023)       0:02:49.572 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.050)       0:02:49.622 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.023)       0:02:49.645 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext3",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "9g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.029)       0:02:49.674 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.024)       0:02:49.699 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.023)       0:02:49.722 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.022)       0:02:49.745 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.021)       0:02:49.766 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.022)       0:02:49.789 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.039)       0:02:49.828 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:10:15 -0500 (0:00:00.020)       0:02:49.848 **** 
changed: [managed-node3] => {
    "actions": [
        {
            "action": "resize device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "resize format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext3"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/mapper/foo-test1",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext3",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "lvm2",
        "xfsprogs",
        "e2fsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext3",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "9g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:10:20 -0500 (0:00:04.784)       0:02:54.633 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:10:20 -0500 (0:00:00.022)       0:02:54.655 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929798.726183,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "54e4dcdb711dd29d857ade258588eb8cc4656237",
        "ctime": 1733929798.723183,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929798.723183,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1394,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:10:20 -0500 (0:00:00.379)       0:02:55.035 **** 
ok: [managed-node3] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:10:21 -0500 (0:00:00.381)       0:02:55.416 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:10:21 -0500 (0:00:00.021)       0:02:55.437 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [
            {
                "action": "resize device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "resize format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext3"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/mapper/foo-test1",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext3",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "lvm2",
            "xfsprogs",
            "e2fsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext3",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "9g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:10:21 -0500 (0:00:00.027)       0:02:55.464 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext3",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "9g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:10:21 -0500 (0:00:00.027)       0:02:55.492 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:10:21 -0500 (0:00:00.025)       0:02:55.518 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:10:21 -0500 (0:00:00.024)       0:02:55.542 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:10:21 -0500 (0:00:00.681)       0:02:56.223 **** 
ok: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext3', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext3",
    "mount_info": {
        "dump": 0,
        "fstype": "ext3",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:10:22 -0500 (0:00:00.414)       0:02:56.637 **** 
skipping: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext3', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "ext3",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:10:22 -0500 (0:00:00.035)       0:02:56.672 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:10:23 -0500 (0:00:00.691)       0:02:57.364 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:10:23 -0500 (0:00:00.394)       0:02:57.759 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:10:23 -0500 (0:00:00.021)       0:02:57.780 **** 
ok: [managed-node3]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:211
Wednesday 11 December 2024  10:10:24 -0500 (0:00:00.813)       0:02:58.593 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node3

TASK [Print out pool information] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Wednesday 11 December 2024  10:10:24 -0500 (0:00:00.044)       0:02:58.637 **** 
ok: [managed-node3] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext3",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "9g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Wednesday 11 December 2024  10:10:24 -0500 (0:00:00.030)       0:02:58.668 **** 
skipping: [managed-node3] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Wednesday 11 December 2024  10:10:24 -0500 (0:00:00.022)       0:02:58.691 **** 
ok: [managed-node3] => {
    "changed": false,
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "ext3",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/foo-test1",
            "size": "9G",
            "type": "lvm",
            "uuid": "c7a0b1fc-e4b0-46bc-b21a-e0e251d60a13"
        },
        "/dev/sda": {
            "fstype": "LVM2_member",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "dfVLB2-vpGU-RqJQ-8Pgk-QQh5-6F0Q-KfgQjh"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Wednesday 11 December 2024  10:10:24 -0500 (0:00:00.387)       0:02:59.078 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003322",
    "end": "2024-12-11 10:10:25.095741",
    "rc": 0,
    "start": "2024-12-11 10:10:25.092419"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/foo-test1 /opt/test1 ext3 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Wednesday 11 December 2024  10:10:25 -0500 (0:00:00.373)       0:02:59.452 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003125",
    "end": "2024-12-11 10:10:25.469380",
    "failed_when_result": false,
    "rc": 0,
    "start": "2024-12-11 10:10:25.466255"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Wednesday 11 December 2024  10:10:25 -0500 (0:00:00.373)       0:02:59.826 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node3

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Wednesday 11 December 2024  10:10:25 -0500 (0:00:00.044)       0:02:59.870 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Wednesday 11 December 2024  10:10:25 -0500 (0:00:00.021)       0:02:59.891 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "vgs",
        "--noheadings",
        "--binary",
        "-o",
        "shared",
        "foo"
    ],
    "delta": "0:00:00.024385",
    "end": "2024-12-11 10:10:25.934093",
    "rc": 0,
    "start": "2024-12-11 10:10:25.909708"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Wednesday 11 December 2024  10:10:25 -0500 (0:00:00.397)       0:03:00.289 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Wednesday 11 December 2024  10:10:26 -0500 (0:00:00.031)       0:03:00.321 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Wednesday 11 December 2024  10:10:26 -0500 (0:00:00.059)       0:03:00.381 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Wednesday 11 December 2024  10:10:26 -0500 (0:00:00.032)       0:03:00.414 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sda",
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Wednesday 11 December 2024  10:10:26 -0500 (0:00:00.400)       0:03:00.814 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Wednesday 11 December 2024  10:10:26 -0500 (0:00:00.026)       0:03:00.841 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Wednesday 11 December 2024  10:10:26 -0500 (0:00:00.028)       0:03:00.869 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Wednesday 11 December 2024  10:10:26 -0500 (0:00:00.029)       0:03:00.899 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Wednesday 11 December 2024  10:10:26 -0500 (0:00:00.027)       0:03:00.926 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Wednesday 11 December 2024  10:10:26 -0500 (0:00:00.032)       0:03:00.958 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Wednesday 11 December 2024  10:10:26 -0500 (0:00:00.026)       0:03:00.985 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Wednesday 11 December 2024  10:10:26 -0500 (0:00:00.049)       0:03:01.035 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.41.57 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.496)       0:03:01.531 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.025)       0:03:01.557 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node3

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.042)       0:03:01.599 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.023)       0:03:01.623 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.024)       0:03:01.648 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.024)       0:03:01.672 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.028)       0:03:01.700 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.034)       0:03:01.735 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.034)       0:03:01.769 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.044)       0:03:01.814 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.044)       0:03:01.858 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.035)       0:03:01.893 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.037)       0:03:01.930 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.037)       0:03:01.968 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node3

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.164)       0:03:02.133 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node3

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.068)       0:03:02.202 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.034)       0:03:02.236 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Wednesday 11 December 2024  10:10:27 -0500 (0:00:00.039)       0:03:02.275 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.046)       0:03:02.322 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.034)       0:03:02.357 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.036)       0:03:02.393 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.068)       0:03:02.462 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.038)       0:03:02.501 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node3

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.072)       0:03:02.574 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node3

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.063)       0:03:02.637 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.033)       0:03:02.671 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.067)       0:03:02.738 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.038)       0:03:02.776 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.032)       0:03:02.809 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.080)       0:03:02.890 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.045)       0:03:02.936 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.040)       0:03:02.976 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node3

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.057)       0:03:03.034 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.029)       0:03:03.063 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.027)       0:03:03.091 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.023)       0:03:03.114 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.021)       0:03:03.135 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.022)       0:03:03.157 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.021)       0:03:03.178 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.020)       0:03:03.199 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node3

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.047)       0:03:03.247 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node3

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Wednesday 11 December 2024  10:10:28 -0500 (0:00:00.045)       0:03:03.293 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.023)       0:03:03.316 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.024)       0:03:03.340 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.023)       0:03:03.364 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.021)       0:03:03.386 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.021)       0:03:03.407 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.021)       0:03:03.429 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.022)       0:03:03.451 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node3

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.067)       0:03:03.518 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.037)       0:03:03.556 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.025)       0:03:03.581 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.027)       0:03:03.608 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.028)       0:03:03.636 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.024)       0:03:03.661 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.025)       0:03:03.686 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.025)       0:03:03.711 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node3

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.046)       0:03:03.757 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.028)       0:03:03.786 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node3

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.158)       0:03:03.945 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.037)       0:03:03.983 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.044)       0:03:04.027 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.025)       0:03:04.052 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.029)       0:03:04.082 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.027)       0:03:04.110 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.027)       0:03:04.137 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.029)       0:03:04.167 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.026)       0:03:04.193 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.024)       0:03:04.217 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.022)       0:03:04.240 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Wednesday 11 December 2024  10:10:29 -0500 (0:00:00.022)       0:03:04.262 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext3 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Wednesday 11 December 2024  10:10:30 -0500 (0:00:00.044)       0:03:04.306 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Wednesday 11 December 2024  10:10:30 -0500 (0:00:00.028)       0:03:04.334 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Wednesday 11 December 2024  10:10:30 -0500 (0:00:00.030)       0:03:04.365 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Wednesday 11 December 2024  10:10:30 -0500 (0:00:00.022)       0:03:04.387 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Wednesday 11 December 2024  10:10:30 -0500 (0:00:00.027)       0:03:04.415 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Wednesday 11 December 2024  10:10:30 -0500 (0:00:00.021)       0:03:04.436 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Wednesday 11 December 2024  10:10:30 -0500 (0:00:00.031)       0:03:04.467 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Wednesday 11 December 2024  10:10:30 -0500 (0:00:00.044)       0:03:04.512 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929820.051107,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1733929820.051107,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 321740,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1733929820.051107,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Wednesday 11 December 2024  10:10:30 -0500 (0:00:00.376)       0:03:04.889 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Wednesday 11 December 2024  10:10:30 -0500 (0:00:00.028)       0:03:04.917 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Wednesday 11 December 2024  10:10:30 -0500 (0:00:00.022)       0:03:04.940 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Wednesday 11 December 2024  10:10:30 -0500 (0:00:00.025)       0:03:04.965 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Wednesday 11 December 2024  10:10:30 -0500 (0:00:00.022)       0:03:04.987 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Wednesday 11 December 2024  10:10:30 -0500 (0:00:00.020)       0:03:05.008 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Wednesday 11 December 2024  10:10:30 -0500 (0:00:00.028)       0:03:05.036 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Wednesday 11 December 2024  10:10:30 -0500 (0:00:00.021)       0:03:05.058 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Wednesday 11 December 2024  10:10:33 -0500 (0:00:02.874)       0:03:07.932 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Wednesday 11 December 2024  10:10:33 -0500 (0:00:00.031)       0:03:07.964 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Wednesday 11 December 2024  10:10:33 -0500 (0:00:00.046)       0:03:08.011 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Wednesday 11 December 2024  10:10:33 -0500 (0:00:00.049)       0:03:08.060 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Wednesday 11 December 2024  10:10:33 -0500 (0:00:00.041)       0:03:08.102 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Wednesday 11 December 2024  10:10:33 -0500 (0:00:00.039)       0:03:08.141 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Wednesday 11 December 2024  10:10:33 -0500 (0:00:00.044)       0:03:08.185 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Wednesday 11 December 2024  10:10:33 -0500 (0:00:00.048)       0:03:08.234 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Wednesday 11 December 2024  10:10:33 -0500 (0:00:00.037)       0:03:08.272 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.042)       0:03:08.314 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.027)       0:03:08.342 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.035)       0:03:08.377 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.025)       0:03:08.403 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.027)       0:03:08.430 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.025)       0:03:08.456 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.023)       0:03:08.480 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.031)       0:03:08.511 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.031)       0:03:08.543 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.031)       0:03:08.574 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.037)       0:03:08.611 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.026)       0:03:08.638 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.031)       0:03:08.669 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.027)       0:03:08.696 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.024)       0:03:08.721 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.022)       0:03:08.743 **** 
ok: [managed-node3] => {
    "bytes": 9663676416,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Wednesday 11 December 2024  10:10:34 -0500 (0:00:00.415)       0:03:09.159 **** 
ok: [managed-node3] => {
    "bytes": 9663676416,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Wednesday 11 December 2024  10:10:35 -0500 (0:00:00.433)       0:03:09.593 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_expected_size": "9663676416"
    },
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Wednesday 11 December 2024  10:10:35 -0500 (0:00:00.046)       0:03:09.639 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "9663676416"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Wednesday 11 December 2024  10:10:35 -0500 (0:00:00.054)       0:03:09.694 **** 
ok: [managed-node3] => {
    "bytes": 10726680821,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Wednesday 11 December 2024  10:10:35 -0500 (0:00:00.532)       0:03:10.227 **** 
skipping: [managed-node3] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Wednesday 11 December 2024  10:10:35 -0500 (0:00:00.040)       0:03:10.267 **** 
skipping: [managed-node3] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.041)       0:03:10.308 **** 
skipping: [managed-node3] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.040)       0:03:10.349 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.032)       0:03:10.381 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.029)       0:03:10.411 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.029)       0:03:10.440 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.027)       0:03:10.467 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.021)       0:03:10.489 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.020)       0:03:10.510 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.021)       0:03:10.532 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.024)       0:03:10.556 **** 
skipping: [managed-node3] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.023)       0:03:10.580 **** 
skipping: [managed-node3] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.020)       0:03:10.601 **** 
skipping: [managed-node3] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.022)       0:03:10.624 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.022)       0:03:10.646 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.026)       0:03:10.673 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.048)       0:03:10.722 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.061)       0:03:10.783 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.045)       0:03:10.828 **** 
ok: [managed-node3] => {
    "storage_test_actual_size": {
        "bytes": 9663676416,
        "changed": false,
        "failed": false,
        "lvm": "9g",
        "parted": "9GiB",
        "size": "9 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.052)       0:03:10.881 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "9663676416"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.041)       0:03:10.922 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Wednesday 11 December 2024  10:10:36 -0500 (0:00:00.081)       0:03:11.003 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.026166",
    "end": "2024-12-11 10:10:37.088107",
    "rc": 0,
    "start": "2024-12-11 10:10:37.061941"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.439)       0:03:11.443 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.028)       0:03:11.472 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.029)       0:03:11.501 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.024)       0:03:11.525 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.024)       0:03:11.549 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.024)       0:03:11.574 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.028)       0:03:11.603 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.029)       0:03:11.632 **** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.034)       0:03:11.666 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Change volume size to 5g] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:214
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.037)       0:03:11.704 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.103)       0:03:11.808 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.042)       0:03:11.850 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.029)       0:03:11.880 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.092)       0:03:11.973 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.041)       0:03:12.015 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.032)       0:03:12.047 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.024)       0:03:12.072 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.026)       0:03:12.098 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.057)       0:03:12.156 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.024)       0:03:12.180 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext3",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "5g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.027)       0:03:12.208 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.024)       0:03:12.232 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.025)       0:03:12.258 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:10:37 -0500 (0:00:00.035)       0:03:12.293 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:10:38 -0500 (0:00:00.031)       0:03:12.324 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:10:38 -0500 (0:00:00.026)       0:03:12.350 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:10:38 -0500 (0:00:00.045)       0:03:12.396 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:10:38 -0500 (0:00:00.024)       0:03:12.421 **** 
changed: [managed-node3] => {
    "actions": [
        {
            "action": "resize format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext3"
        },
        {
            "action": "resize device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/mapper/foo-test1",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext3",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "lvm2",
        "e2fsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext3",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:10:43 -0500 (0:00:05.736)       0:03:18.157 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:10:43 -0500 (0:00:00.034)       0:03:18.192 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929798.726183,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "54e4dcdb711dd29d857ade258588eb8cc4656237",
        "ctime": 1733929798.723183,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929798.723183,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1394,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:10:44 -0500 (0:00:00.432)       0:03:18.625 **** 
ok: [managed-node3] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:10:44 -0500 (0:00:00.456)       0:03:19.082 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:10:44 -0500 (0:00:00.029)       0:03:19.112 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [
            {
                "action": "resize format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext3"
            },
            {
                "action": "resize device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/mapper/foo-test1",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext3",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "lvm2",
            "e2fsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext3",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:10:44 -0500 (0:00:00.045)       0:03:19.157 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext3",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:10:44 -0500 (0:00:00.041)       0:03:19.198 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:10:44 -0500 (0:00:00.037)       0:03:19.235 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:10:44 -0500 (0:00:00.032)       0:03:19.268 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:10:45 -0500 (0:00:00.688)       0:03:19.956 **** 
changed: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext3', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext3",
    "mount_info": {
        "dump": 0,
        "fstype": "ext3",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:10:46 -0500 (0:00:00.408)       0:03:20.364 **** 
skipping: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext3', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "ext3",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:10:46 -0500 (0:00:00.029)       0:03:20.394 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:10:46 -0500 (0:00:00.641)       0:03:21.036 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:10:47 -0500 (0:00:00.385)       0:03:21.422 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:10:47 -0500 (0:00:00.021)       0:03:21.443 **** 
ok: [managed-node3]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:228
Wednesday 11 December 2024  10:10:47 -0500 (0:00:00.765)       0:03:22.208 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node3

TASK [Print out pool information] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Wednesday 11 December 2024  10:10:47 -0500 (0:00:00.046)       0:03:22.255 **** 
ok: [managed-node3] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext3",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Wednesday 11 December 2024  10:10:47 -0500 (0:00:00.028)       0:03:22.283 **** 
skipping: [managed-node3] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Wednesday 11 December 2024  10:10:48 -0500 (0:00:00.021)       0:03:22.304 **** 
ok: [managed-node3] => {
    "changed": false,
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "ext3",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/foo-test1",
            "size": "5G",
            "type": "lvm",
            "uuid": "c7a0b1fc-e4b0-46bc-b21a-e0e251d60a13"
        },
        "/dev/sda": {
            "fstype": "LVM2_member",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "dfVLB2-vpGU-RqJQ-8Pgk-QQh5-6F0Q-KfgQjh"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Wednesday 11 December 2024  10:10:48 -0500 (0:00:00.373)       0:03:22.678 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002646",
    "end": "2024-12-11 10:10:48.698095",
    "rc": 0,
    "start": "2024-12-11 10:10:48.695449"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/foo-test1 /opt/test1 ext3 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Wednesday 11 December 2024  10:10:48 -0500 (0:00:00.375)       0:03:23.054 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002742",
    "end": "2024-12-11 10:10:49.073009",
    "failed_when_result": false,
    "rc": 0,
    "start": "2024-12-11 10:10:49.070267"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Wednesday 11 December 2024  10:10:49 -0500 (0:00:00.412)       0:03:23.466 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node3

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Wednesday 11 December 2024  10:10:49 -0500 (0:00:00.043)       0:03:23.509 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Wednesday 11 December 2024  10:10:49 -0500 (0:00:00.020)       0:03:23.530 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "vgs",
        "--noheadings",
        "--binary",
        "-o",
        "shared",
        "foo"
    ],
    "delta": "0:00:00.024496",
    "end": "2024-12-11 10:10:49.574739",
    "rc": 0,
    "start": "2024-12-11 10:10:49.550243"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Wednesday 11 December 2024  10:10:49 -0500 (0:00:00.400)       0:03:23.930 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Wednesday 11 December 2024  10:10:49 -0500 (0:00:00.032)       0:03:23.963 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Wednesday 11 December 2024  10:10:49 -0500 (0:00:00.048)       0:03:24.011 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Wednesday 11 December 2024  10:10:49 -0500 (0:00:00.031)       0:03:24.042 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sda",
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Wednesday 11 December 2024  10:10:50 -0500 (0:00:00.400)       0:03:24.443 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Wednesday 11 December 2024  10:10:50 -0500 (0:00:00.024)       0:03:24.467 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Wednesday 11 December 2024  10:10:50 -0500 (0:00:00.029)       0:03:24.496 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Wednesday 11 December 2024  10:10:50 -0500 (0:00:00.030)       0:03:24.527 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Wednesday 11 December 2024  10:10:50 -0500 (0:00:00.036)       0:03:24.564 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Wednesday 11 December 2024  10:10:50 -0500 (0:00:00.036)       0:03:24.600 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Wednesday 11 December 2024  10:10:50 -0500 (0:00:00.025)       0:03:24.625 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Wednesday 11 December 2024  10:10:50 -0500 (0:00:00.038)       0:03:24.664 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.41.57 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Wednesday 11 December 2024  10:10:50 -0500 (0:00:00.454)       0:03:25.119 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Wednesday 11 December 2024  10:10:50 -0500 (0:00:00.028)       0:03:25.147 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node3

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Wednesday 11 December 2024  10:10:50 -0500 (0:00:00.073)       0:03:25.221 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Wednesday 11 December 2024  10:10:50 -0500 (0:00:00.027)       0:03:25.248 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Wednesday 11 December 2024  10:10:50 -0500 (0:00:00.031)       0:03:25.280 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.022)       0:03:25.303 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.021)       0:03:25.325 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.021)       0:03:25.347 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.022)       0:03:25.369 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.021)       0:03:25.390 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.023)       0:03:25.414 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.021)       0:03:25.435 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.025)       0:03:25.461 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.024)       0:03:25.485 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node3

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.066)       0:03:25.551 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node3

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.051)       0:03:25.603 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.023)       0:03:25.627 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.032)       0:03:25.659 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.027)       0:03:25.686 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.026)       0:03:25.713 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.025)       0:03:25.738 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.022)       0:03:25.761 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.023)       0:03:25.784 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node3

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.045)       0:03:25.829 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node3

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.044)       0:03:25.873 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.023)       0:03:25.896 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.022)       0:03:25.919 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.024)       0:03:25.944 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.058)       0:03:26.002 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.047)       0:03:26.050 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.027)       0:03:26.077 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.027)       0:03:26.104 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node3

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.044)       0:03:26.149 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.030)       0:03:26.179 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.042)       0:03:26.222 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.037)       0:03:26.259 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Wednesday 11 December 2024  10:10:51 -0500 (0:00:00.032)       0:03:26.291 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.027)       0:03:26.319 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.024)       0:03:26.343 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.027)       0:03:26.371 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node3

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.057)       0:03:26.428 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node3

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.046)       0:03:26.474 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.023)       0:03:26.497 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.022)       0:03:26.520 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.024)       0:03:26.545 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.023)       0:03:26.568 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.022)       0:03:26.591 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.025)       0:03:26.616 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.024)       0:03:26.641 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node3

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.078)       0:03:26.719 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.033)       0:03:26.753 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.025)       0:03:26.778 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.027)       0:03:26.806 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.023)       0:03:26.829 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.029)       0:03:26.858 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.028)       0:03:26.887 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.023)       0:03:26.911 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node3

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.047)       0:03:26.959 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.030)       0:03:26.989 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node3

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.115)       0:03:27.105 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.025)       0:03:27.131 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.029)       0:03:27.160 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.025)       0:03:27.186 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.027)       0:03:27.213 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.021)       0:03:27.235 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.022)       0:03:27.258 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Wednesday 11 December 2024  10:10:52 -0500 (0:00:00.022)       0:03:27.280 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.034)       0:03:27.315 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.024)       0:03:27.339 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.023)       0:03:27.362 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.023)       0:03:27.386 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext3 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.056)       0:03:27.442 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.035)       0:03:27.478 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.032)       0:03:27.510 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.081)       0:03:27.591 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.030)       0:03:27.622 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.025)       0:03:27.647 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.031)       0:03:27.679 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.033)       0:03:27.713 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929843.6830297,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1733929843.6830297,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 321740,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1733929843.6830297,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.389)       0:03:28.103 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.029)       0:03:28.132 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.023)       0:03:28.156 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.028)       0:03:28.185 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.025)       0:03:28.210 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.024)       0:03:28.235 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.027)       0:03:28.262 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Wednesday 11 December 2024  10:10:53 -0500 (0:00:00.022)       0:03:28.284 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Wednesday 11 December 2024  10:10:56 -0500 (0:00:02.881)       0:03:31.166 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Wednesday 11 December 2024  10:10:56 -0500 (0:00:00.034)       0:03:31.201 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Wednesday 11 December 2024  10:10:56 -0500 (0:00:00.033)       0:03:31.235 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Wednesday 11 December 2024  10:10:56 -0500 (0:00:00.049)       0:03:31.285 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.039)       0:03:31.324 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.035)       0:03:31.359 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.033)       0:03:31.392 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.041)       0:03:31.434 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.048)       0:03:31.482 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.041)       0:03:31.523 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.039)       0:03:31.563 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.035)       0:03:31.599 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.032)       0:03:31.632 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.036)       0:03:31.669 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.033)       0:03:31.702 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.036)       0:03:31.739 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.035)       0:03:31.774 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.025)       0:03:31.800 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.025)       0:03:31.826 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.028)       0:03:31.855 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.025)       0:03:31.880 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.028)       0:03:31.908 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.026)       0:03:31.935 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.025)       0:03:31.961 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Wednesday 11 December 2024  10:10:57 -0500 (0:00:00.022)       0:03:31.984 **** 
ok: [managed-node3] => {
    "bytes": 5368709120,
    "changed": false,
    "lvm": "5g",
    "parted": "5GiB",
    "size": "5 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Wednesday 11 December 2024  10:10:58 -0500 (0:00:00.402)       0:03:32.386 **** 
ok: [managed-node3] => {
    "bytes": 5368709120,
    "changed": false,
    "lvm": "5g",
    "parted": "5GiB",
    "size": "5 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Wednesday 11 December 2024  10:10:58 -0500 (0:00:00.415)       0:03:32.802 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_expected_size": "5368709120"
    },
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Wednesday 11 December 2024  10:10:58 -0500 (0:00:00.030)       0:03:32.832 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Wednesday 11 December 2024  10:10:58 -0500 (0:00:00.024)       0:03:32.857 **** 
ok: [managed-node3] => {
    "bytes": 10726680821,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Wednesday 11 December 2024  10:10:58 -0500 (0:00:00.385)       0:03:33.243 **** 
skipping: [managed-node3] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Wednesday 11 December 2024  10:10:58 -0500 (0:00:00.027)       0:03:33.270 **** 
skipping: [managed-node3] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Wednesday 11 December 2024  10:10:58 -0500 (0:00:00.027)       0:03:33.297 **** 
skipping: [managed-node3] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.026)       0:03:33.324 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.025)       0:03:33.349 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.021)       0:03:33.371 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.022)       0:03:33.393 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.021)       0:03:33.415 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.026)       0:03:33.441 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.024)       0:03:33.466 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.032)       0:03:33.499 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.036)       0:03:33.535 **** 
skipping: [managed-node3] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.031)       0:03:33.567 **** 
skipping: [managed-node3] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.027)       0:03:33.594 **** 
skipping: [managed-node3] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.027)       0:03:33.622 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.027)       0:03:33.650 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.028)       0:03:33.679 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.025)       0:03:33.704 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.026)       0:03:33.730 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.022)       0:03:33.753 **** 
ok: [managed-node3] => {
    "storage_test_actual_size": {
        "bytes": 5368709120,
        "changed": false,
        "failed": false,
        "lvm": "5g",
        "parted": "5GiB",
        "size": "5 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.028)       0:03:33.781 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.024)       0:03:33.806 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.032)       0:03:33.838 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.024762",
    "end": "2024-12-11 10:10:59.902100",
    "rc": 0,
    "start": "2024-12-11 10:10:59.877338"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.419)       0:03:34.257 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Wednesday 11 December 2024  10:10:59 -0500 (0:00:00.029)       0:03:34.286 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.029)       0:03:34.316 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.024)       0:03:34.341 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.076)       0:03:34.417 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.035)       0:03:34.453 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.034)       0:03:34.487 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.025)       0:03:34.512 **** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.022)       0:03:34.534 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Clean up] ****************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:231
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.027)       0:03:34.562 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.096)       0:03:34.659 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.033)       0:03:34.693 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.026)       0:03:34.719 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.057)       0:03:34.777 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.023)       0:03:34.800 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.021)       0:03:34.821 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.025)       0:03:34.847 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.021)       0:03:34.868 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.049)       0:03:34.917 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.024)       0:03:34.942 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "state": "absent",
            "volumes": [
                {
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "5g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.028)       0:03:34.970 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.024)       0:03:34.994 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.021)       0:03:35.016 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.022)       0:03:35.039 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.022)       0:03:35.061 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.025)       0:03:35.086 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.039)       0:03:35.126 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:11:00 -0500 (0:00:00.030)       0:03:35.157 **** 
changed: [managed-node3] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext3"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sda",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1"
    ],
    "mounts": [
        {
            "fstype": "ext3",
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "absent",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext3",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:11:05 -0500 (0:00:05.082)       0:03:40.239 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:11:05 -0500 (0:00:00.024)       0:03:40.263 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929798.726183,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "54e4dcdb711dd29d857ade258588eb8cc4656237",
        "ctime": 1733929798.723183,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929798.723183,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1394,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:11:06 -0500 (0:00:00.405)       0:03:40.669 **** 
ok: [managed-node3] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:11:06 -0500 (0:00:00.447)       0:03:41.117 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:11:06 -0500 (0:00:00.019)       0:03:41.136 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext3"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sda",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1"
        ],
        "mounts": [
            {
                "fstype": "ext3",
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "absent",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext3",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:11:06 -0500 (0:00:00.027)       0:03:41.163 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "absent",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext3",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:11:06 -0500 (0:00:00.026)       0:03:41.190 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:11:06 -0500 (0:00:00.023)       0:03:41.213 **** 
changed: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'ext3'}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext3",
    "mount_info": {
        "fstype": "ext3",
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:11:07 -0500 (0:00:00.409)       0:03:41.623 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:11:07 -0500 (0:00:00.658)       0:03:42.282 **** 

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:11:08 -0500 (0:00:00.023)       0:03:42.306 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:11:08 -0500 (0:00:00.024)       0:03:42.330 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:11:08 -0500 (0:00:00.656)       0:03:42.986 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:11:09 -0500 (0:00:00.392)       0:03:43.379 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:11:09 -0500 (0:00:00.030)       0:03:43.409 **** 
ok: [managed-node3]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:244
Wednesday 11 December 2024  10:11:09 -0500 (0:00:00.801)       0:03:44.211 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node3

TASK [Print out pool information] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Wednesday 11 December 2024  10:11:09 -0500 (0:00:00.078)       0:03:44.290 **** 
ok: [managed-node3] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "absent",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext3",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Wednesday 11 December 2024  10:11:10 -0500 (0:00:00.044)       0:03:44.334 **** 
skipping: [managed-node3] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Wednesday 11 December 2024  10:11:10 -0500 (0:00:00.040)       0:03:44.375 **** 
ok: [managed-node3] => {
    "changed": false,
    "info": {
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Wednesday 11 December 2024  10:11:10 -0500 (0:00:00.515)       0:03:44.890 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002632",
    "end": "2024-12-11 10:11:10.953283",
    "rc": 0,
    "start": "2024-12-11 10:11:10.950651"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Wednesday 11 December 2024  10:11:11 -0500 (0:00:00.478)       0:03:45.369 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002656",
    "end": "2024-12-11 10:11:11.483476",
    "failed_when_result": false,
    "rc": 0,
    "start": "2024-12-11 10:11:11.480820"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Wednesday 11 December 2024  10:11:11 -0500 (0:00:00.487)       0:03:45.857 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node3

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Wednesday 11 December 2024  10:11:11 -0500 (0:00:00.074)       0:03:45.931 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Wednesday 11 December 2024  10:11:11 -0500 (0:00:00.047)       0:03:45.979 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Wednesday 11 December 2024  10:11:11 -0500 (0:00:00.040)       0:03:46.019 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Wednesday 11 December 2024  10:11:11 -0500 (0:00:00.058)       0:03:46.078 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Wednesday 11 December 2024  10:11:11 -0500 (0:00:00.075)       0:03:46.153 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "0",
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Wednesday 11 December 2024  10:11:11 -0500 (0:00:00.046)       0:03:46.200 **** 

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Wednesday 11 December 2024  10:11:11 -0500 (0:00:00.021)       0:03:46.222 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": "0"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Wednesday 11 December 2024  10:11:11 -0500 (0:00:00.032)       0:03:46.254 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": []
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Wednesday 11 December 2024  10:11:11 -0500 (0:00:00.028)       0:03:46.283 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Wednesday 11 December 2024  10:11:12 -0500 (0:00:00.030)       0:03:46.313 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Wednesday 11 December 2024  10:11:12 -0500 (0:00:00.026)       0:03:46.340 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Wednesday 11 December 2024  10:11:12 -0500 (0:00:00.026)       0:03:46.367 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Wednesday 11 December 2024  10:11:12 -0500 (0:00:00.025)       0:03:46.393 **** 

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Wednesday 11 December 2024  10:11:12 -0500 (0:00:00.019)       0:03:46.413 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.41.57 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Wednesday 11 December 2024  10:11:12 -0500 (0:00:00.440)       0:03:46.854 **** 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Wednesday 11 December 2024  10:11:12 -0500 (0:00:00.034)       0:03:46.888 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node3

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Wednesday 11 December 2024  10:11:12 -0500 (0:00:00.069)       0:03:46.958 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Wednesday 11 December 2024  10:11:12 -0500 (0:00:00.032)       0:03:46.991 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Wednesday 11 December 2024  10:11:12 -0500 (0:00:00.033)       0:03:47.024 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Wednesday 11 December 2024  10:11:12 -0500 (0:00:00.033)       0:03:47.057 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Wednesday 11 December 2024  10:11:12 -0500 (0:00:00.036)       0:03:47.094 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Wednesday 11 December 2024  10:11:12 -0500 (0:00:00.109)       0:03:47.203 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Wednesday 11 December 2024  10:11:12 -0500 (0:00:00.029)       0:03:47.233 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Wednesday 11 December 2024  10:11:12 -0500 (0:00:00.040)       0:03:47.273 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.033)       0:03:47.307 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.034)       0:03:47.341 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.032)       0:03:47.375 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.035)       0:03:47.410 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node3

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.071)       0:03:47.482 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node3

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.064)       0:03:47.546 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.039)       0:03:47.585 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.036)       0:03:47.622 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.033)       0:03:47.655 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.032)       0:03:47.687 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.033)       0:03:47.721 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.033)       0:03:47.754 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.037)       0:03:47.791 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node3

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.069)       0:03:47.861 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node3

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.058)       0:03:47.919 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.024)       0:03:47.944 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.024)       0:03:47.969 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.026)       0:03:47.995 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.022)       0:03:48.018 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.060)       0:03:48.078 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.027)       0:03:48.106 **** 

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.018)       0:03:48.125 **** 

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.027)       0:03:48.153 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.023)       0:03:48.176 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node3

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.048)       0:03:48.225 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node3

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.048)       0:03:48.273 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Wednesday 11 December 2024  10:11:13 -0500 (0:00:00.023)       0:03:48.297 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.036)       0:03:48.333 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.033)       0:03:48.366 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.030)       0:03:48.397 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.032)       0:03:48.430 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.033)       0:03:48.463 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.033)       0:03:48.497 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node3

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.081)       0:03:48.579 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.034)       0:03:48.613 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.035)       0:03:48.649 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.034)       0:03:48.684 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.029)       0:03:48.713 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.026)       0:03:48.740 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.022)       0:03:48.763 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.025)       0:03:48.788 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node3

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.048)       0:03:48.837 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": false,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.032)       0:03:48.869 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node3

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.150)       0:03:49.020 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.025)       0:03:49.046 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.028)       0:03:49.075 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.021)       0:03:49.096 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.018)       0:03:49.115 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.021)       0:03:49.137 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.021)       0:03:49.158 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.032)       0:03:49.191 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.020)       0:03:49.211 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.021)       0:03:49.232 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.022)       0:03:49.255 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Wednesday 11 December 2024  10:11:14 -0500 (0:00:00.020)       0:03:49.275 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "0",
        "storage_test_fstab_expected_mount_options_matches": "0",
        "storage_test_fstab_expected_mount_point_matches": "0",
        "storage_test_fstab_id_matches": [],
        "storage_test_fstab_mount_options_matches": [],
        "storage_test_fstab_mount_point_matches": []
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Wednesday 11 December 2024  10:11:15 -0500 (0:00:00.042)       0:03:49.318 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Wednesday 11 December 2024  10:11:15 -0500 (0:00:00.020)       0:03:49.338 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Wednesday 11 December 2024  10:11:15 -0500 (0:00:00.026)       0:03:49.365 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Wednesday 11 December 2024  10:11:15 -0500 (0:00:00.023)       0:03:49.388 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Wednesday 11 December 2024  10:11:15 -0500 (0:00:00.026)       0:03:49.415 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Wednesday 11 December 2024  10:11:15 -0500 (0:00:00.020)       0:03:49.435 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Wednesday 11 December 2024  10:11:15 -0500 (0:00:00.022)       0:03:49.458 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Wednesday 11 December 2024  10:11:15 -0500 (0:00:00.023)       0:03:49.482 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "exists": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Wednesday 11 December 2024  10:11:15 -0500 (0:00:00.368)       0:03:49.850 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Wednesday 11 December 2024  10:11:15 -0500 (0:00:00.022)       0:03:49.872 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Wednesday 11 December 2024  10:11:15 -0500 (0:00:00.028)       0:03:49.901 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Wednesday 11 December 2024  10:11:15 -0500 (0:00:00.017)       0:03:49.918 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Wednesday 11 December 2024  10:11:15 -0500 (0:00:00.024)       0:03:49.943 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Wednesday 11 December 2024  10:11:15 -0500 (0:00:00.022)       0:03:49.966 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Wednesday 11 December 2024  10:11:15 -0500 (0:00:00.018)       0:03:49.984 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Wednesday 11 December 2024  10:11:15 -0500 (0:00:00.021)       0:03:50.006 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Wednesday 11 December 2024  10:11:18 -0500 (0:00:02.829)       0:03:52.835 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Wednesday 11 December 2024  10:11:18 -0500 (0:00:00.034)       0:03:52.869 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Wednesday 11 December 2024  10:11:18 -0500 (0:00:00.036)       0:03:52.906 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Wednesday 11 December 2024  10:11:18 -0500 (0:00:00.028)       0:03:52.935 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Wednesday 11 December 2024  10:11:18 -0500 (0:00:00.032)       0:03:52.967 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Wednesday 11 December 2024  10:11:18 -0500 (0:00:00.031)       0:03:52.999 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Wednesday 11 December 2024  10:11:18 -0500 (0:00:00.027)       0:03:53.027 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Wednesday 11 December 2024  10:11:18 -0500 (0:00:00.025)       0:03:53.052 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Wednesday 11 December 2024  10:11:18 -0500 (0:00:00.025)       0:03:53.078 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Wednesday 11 December 2024  10:11:18 -0500 (0:00:00.042)       0:03:53.120 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Wednesday 11 December 2024  10:11:18 -0500 (0:00:00.039)       0:03:53.160 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Wednesday 11 December 2024  10:11:18 -0500 (0:00:00.033)       0:03:53.193 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Wednesday 11 December 2024  10:11:18 -0500 (0:00:00.031)       0:03:53.225 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Wednesday 11 December 2024  10:11:18 -0500 (0:00:00.032)       0:03:53.258 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.052)       0:03:53.311 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.049)       0:03:53.360 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.032)       0:03:53.393 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.033)       0:03:53.427 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.036)       0:03:53.464 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.033)       0:03:53.498 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.033)       0:03:53.531 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.036)       0:03:53.567 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.035)       0:03:53.603 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.030)       0:03:53.633 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.034)       0:03:53.667 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.033)       0:03:53.701 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.032)       0:03:53.733 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.038)       0:03:53.771 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.036)       0:03:53.808 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.034)       0:03:53.842 **** 
skipping: [managed-node3] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.024)       0:03:53.867 **** 
skipping: [managed-node3] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.025)       0:03:53.893 **** 
skipping: [managed-node3] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.022)       0:03:53.915 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.028)       0:03:53.944 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.076)       0:03:54.020 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.023)       0:03:54.044 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.021)       0:03:54.066 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.022)       0:03:54.089 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.021)       0:03:54.110 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.021)       0:03:54.132 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.023)       0:03:54.156 **** 
skipping: [managed-node3] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.022)       0:03:54.178 **** 
skipping: [managed-node3] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.022)       0:03:54.200 **** 
skipping: [managed-node3] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.021)       0:03:54.222 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.021)       0:03:54.244 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.021)       0:03:54.265 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Wednesday 11 December 2024  10:11:19 -0500 (0:00:00.022)       0:03:54.288 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.021)       0:03:54.309 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.021)       0:03:54.330 **** 
ok: [managed-node3] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.038)       0:03:54.369 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.024)       0:03:54.394 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.021)       0:03:54.415 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.023)       0:03:54.439 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.022)       0:03:54.462 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.022)       0:03:54.484 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.022)       0:03:54.507 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.023)       0:03:54.530 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.022)       0:03:54.552 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.024)       0:03:54.576 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.021)       0:03:54.598 **** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.019)       0:03:54.618 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Create a LVM logical volume with for ext2 FS size 5g] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:249
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.022)       0:03:54.640 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.075)       0:03:54.716 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.031)       0:03:54.747 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.025)       0:03:54.773 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.057)       0:03:54.831 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.022)       0:03:54.853 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.022)       0:03:54.876 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.020)       0:03:54.896 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.022)       0:03:54.918 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.048)       0:03:54.967 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.021)       0:03:54.988 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext2",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "5g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.028)       0:03:55.017 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.024)       0:03:55.041 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.020)       0:03:55.062 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.021)       0:03:55.084 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.022)       0:03:55.106 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.022)       0:03:55.129 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.034)       0:03:55.164 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:11:20 -0500 (0:00:00.020)       0:03:55.184 **** 
changed: [managed-node3] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/sda",
            "fs_type": "lvmpv"
        },
        {
            "action": "create device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "create device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext2"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1",
        "/dev/mapper/foo-test1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext2",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "lvm2",
        "e2fsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext2",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:11:25 -0500 (0:00:04.886)       0:04:00.070 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:11:25 -0500 (0:00:00.033)       0:04:00.104 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929867.800951,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "ab8070345774adad92683e9645714452be7be474",
        "ctime": 1733929867.247953,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929867.247953,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1343,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:11:26 -0500 (0:00:00.445)       0:04:00.550 **** 
ok: [managed-node3] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:11:26 -0500 (0:00:00.388)       0:04:00.938 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:11:26 -0500 (0:00:00.019)       0:04:00.958 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/sda",
                "fs_type": "lvmpv"
            },
            {
                "action": "create device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "create device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext2"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1",
            "/dev/mapper/foo-test1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext2",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "lvm2",
            "e2fsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext2",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:11:26 -0500 (0:00:00.028)       0:04:00.986 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext2",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:11:26 -0500 (0:00:00.026)       0:04:01.013 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:11:26 -0500 (0:00:00.024)       0:04:01.038 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:11:26 -0500 (0:00:00.024)       0:04:01.063 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:11:27 -0500 (0:00:00.713)       0:04:01.776 **** 
changed: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext2', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext2",
    "mount_info": {
        "dump": 0,
        "fstype": "ext2",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:11:27 -0500 (0:00:00.445)       0:04:02.221 **** 
skipping: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext2', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "ext2",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:11:27 -0500 (0:00:00.038)       0:04:02.259 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:11:28 -0500 (0:00:00.664)       0:04:02.924 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:11:29 -0500 (0:00:00.513)       0:04:03.437 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:11:29 -0500 (0:00:00.029)       0:04:03.467 **** 
ok: [managed-node3]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:265
Wednesday 11 December 2024  10:11:30 -0500 (0:00:00.880)       0:04:04.347 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node3

TASK [Print out pool information] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Wednesday 11 December 2024  10:11:30 -0500 (0:00:00.079)       0:04:04.427 **** 
ok: [managed-node3] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext2",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Wednesday 11 December 2024  10:11:30 -0500 (0:00:00.040)       0:04:04.467 **** 
skipping: [managed-node3] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Wednesday 11 December 2024  10:11:30 -0500 (0:00:00.025)       0:04:04.493 **** 
ok: [managed-node3] => {
    "changed": false,
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "ext2",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/foo-test1",
            "size": "5G",
            "type": "lvm",
            "uuid": "78de8819-280f-4432-9837-123353bb821c"
        },
        "/dev/sda": {
            "fstype": "LVM2_member",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "VKmjYR-rX7d-wXez-nwPP-hiIt-jeKd-MyNwyV"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Wednesday 11 December 2024  10:11:30 -0500 (0:00:00.404)       0:04:04.897 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003237",
    "end": "2024-12-11 10:11:30.929749",
    "rc": 0,
    "start": "2024-12-11 10:11:30.926512"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/foo-test1 /opt/test1 ext2 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Wednesday 11 December 2024  10:11:30 -0500 (0:00:00.388)       0:04:05.285 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002671",
    "end": "2024-12-11 10:11:31.304146",
    "failed_when_result": false,
    "rc": 0,
    "start": "2024-12-11 10:11:31.301475"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Wednesday 11 December 2024  10:11:31 -0500 (0:00:00.383)       0:04:05.668 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node3

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Wednesday 11 December 2024  10:11:31 -0500 (0:00:00.044)       0:04:05.712 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Wednesday 11 December 2024  10:11:31 -0500 (0:00:00.021)       0:04:05.734 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "vgs",
        "--noheadings",
        "--binary",
        "-o",
        "shared",
        "foo"
    ],
    "delta": "0:00:00.022727",
    "end": "2024-12-11 10:11:31.780952",
    "rc": 0,
    "start": "2024-12-11 10:11:31.758225"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Wednesday 11 December 2024  10:11:31 -0500 (0:00:00.402)       0:04:06.136 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Wednesday 11 December 2024  10:11:31 -0500 (0:00:00.032)       0:04:06.168 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Wednesday 11 December 2024  10:11:31 -0500 (0:00:00.050)       0:04:06.219 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Wednesday 11 December 2024  10:11:31 -0500 (0:00:00.030)       0:04:06.250 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sda",
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Wednesday 11 December 2024  10:11:32 -0500 (0:00:00.378)       0:04:06.628 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Wednesday 11 December 2024  10:11:32 -0500 (0:00:00.026)       0:04:06.655 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Wednesday 11 December 2024  10:11:32 -0500 (0:00:00.038)       0:04:06.694 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Wednesday 11 December 2024  10:11:32 -0500 (0:00:00.028)       0:04:06.723 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Wednesday 11 December 2024  10:11:32 -0500 (0:00:00.026)       0:04:06.749 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Wednesday 11 December 2024  10:11:32 -0500 (0:00:00.026)       0:04:06.775 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Wednesday 11 December 2024  10:11:32 -0500 (0:00:00.021)       0:04:06.797 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Wednesday 11 December 2024  10:11:32 -0500 (0:00:00.033)       0:04:06.831 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.41.57 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Wednesday 11 December 2024  10:11:32 -0500 (0:00:00.438)       0:04:07.269 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.035)       0:04:07.305 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node3

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.061)       0:04:07.367 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.031)       0:04:07.398 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.033)       0:04:07.432 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.032)       0:04:07.464 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.032)       0:04:07.497 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.034)       0:04:07.531 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.031)       0:04:07.563 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.030)       0:04:07.594 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.027)       0:04:07.622 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.023)       0:04:07.645 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.026)       0:04:07.672 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.025)       0:04:07.698 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node3

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.075)       0:04:07.773 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node3

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.056)       0:04:07.830 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.035)       0:04:07.865 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.032)       0:04:07.897 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.035)       0:04:07.933 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.034)       0:04:07.967 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.037)       0:04:08.005 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.079)       0:04:08.084 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.055)       0:04:08.140 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node3

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Wednesday 11 December 2024  10:11:33 -0500 (0:00:00.109)       0:04:08.249 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node3

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.062)       0:04:08.312 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.033)       0:04:08.346 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.032)       0:04:08.378 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.113)       0:04:08.492 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.041)       0:04:08.533 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.081)       0:04:08.615 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.044)       0:04:08.659 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.053)       0:04:08.712 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node3

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.075)       0:04:08.788 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.051)       0:04:08.840 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.040)       0:04:08.880 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.051)       0:04:08.932 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.086)       0:04:09.019 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.044)       0:04:09.064 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.030)       0:04:09.095 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.053)       0:04:09.148 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node3

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Wednesday 11 December 2024  10:11:34 -0500 (0:00:00.134)       0:04:09.283 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node3

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.074)       0:04:09.358 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.051)       0:04:09.410 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.035)       0:04:09.445 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.035)       0:04:09.480 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.036)       0:04:09.517 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.031)       0:04:09.549 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.033)       0:04:09.583 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.039)       0:04:09.622 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node3

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.087)       0:04:09.710 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.033)       0:04:09.744 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.033)       0:04:09.777 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.033)       0:04:09.810 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.048)       0:04:09.859 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.067)       0:04:09.927 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.038)       0:04:09.966 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.048)       0:04:10.014 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node3

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.076)       0:04:10.090 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Wednesday 11 December 2024  10:11:35 -0500 (0:00:00.051)       0:04:10.142 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node3

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.212)       0:04:10.355 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.039)       0:04:10.394 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.087)       0:04:10.482 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.041)       0:04:10.524 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.071)       0:04:10.595 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.044)       0:04:10.639 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.040)       0:04:10.680 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.035)       0:04:10.715 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.035)       0:04:10.751 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.044)       0:04:10.796 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.036)       0:04:10.832 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.035)       0:04:10.867 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext2 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.083)       0:04:10.951 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.061)       0:04:11.013 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.040)       0:04:11.053 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.031)       0:04:11.084 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.139)       0:04:11.223 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Wednesday 11 December 2024  10:11:36 -0500 (0:00:00.047)       0:04:11.271 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Wednesday 11 December 2024  10:11:37 -0500 (0:00:00.073)       0:04:11.345 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Wednesday 11 December 2024  10:11:37 -0500 (0:00:00.050)       0:04:11.395 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929885.6308928,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1733929885.6308928,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 347287,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1733929885.6308928,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Wednesday 11 December 2024  10:11:37 -0500 (0:00:00.662)       0:04:12.058 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Wednesday 11 December 2024  10:11:37 -0500 (0:00:00.073)       0:04:12.132 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Wednesday 11 December 2024  10:11:37 -0500 (0:00:00.043)       0:04:12.176 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Wednesday 11 December 2024  10:11:37 -0500 (0:00:00.059)       0:04:12.235 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Wednesday 11 December 2024  10:11:37 -0500 (0:00:00.035)       0:04:12.270 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Wednesday 11 December 2024  10:11:38 -0500 (0:00:00.032)       0:04:12.303 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Wednesday 11 December 2024  10:11:38 -0500 (0:00:00.042)       0:04:12.345 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Wednesday 11 December 2024  10:11:38 -0500 (0:00:00.034)       0:04:12.380 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Wednesday 11 December 2024  10:11:41 -0500 (0:00:02.921)       0:04:15.301 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.030)       0:04:15.332 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.033)       0:04:15.366 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.045)       0:04:15.411 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.034)       0:04:15.445 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.034)       0:04:15.480 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.033)       0:04:15.513 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.035)       0:04:15.549 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.031)       0:04:15.580 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.030)       0:04:15.611 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.031)       0:04:15.643 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.034)       0:04:15.677 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.037)       0:04:15.715 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.041)       0:04:15.757 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.025)       0:04:15.782 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.027)       0:04:15.810 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.027)       0:04:15.838 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.031)       0:04:15.870 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.022)       0:04:15.892 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.022)       0:04:15.914 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.021)       0:04:15.936 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.022)       0:04:15.958 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.026)       0:04:15.984 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.021)       0:04:16.006 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Wednesday 11 December 2024  10:11:41 -0500 (0:00:00.022)       0:04:16.028 **** 
ok: [managed-node3] => {
    "bytes": 5368709120,
    "changed": false,
    "lvm": "5g",
    "parted": "5GiB",
    "size": "5 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Wednesday 11 December 2024  10:11:42 -0500 (0:00:00.458)       0:04:16.487 **** 
ok: [managed-node3] => {
    "bytes": 5368709120,
    "changed": false,
    "lvm": "5g",
    "parted": "5GiB",
    "size": "5 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Wednesday 11 December 2024  10:11:42 -0500 (0:00:00.392)       0:04:16.880 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_expected_size": "5368709120"
    },
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Wednesday 11 December 2024  10:11:42 -0500 (0:00:00.030)       0:04:16.910 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Wednesday 11 December 2024  10:11:42 -0500 (0:00:00.024)       0:04:16.935 **** 
ok: [managed-node3] => {
    "bytes": 10726680821,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.379)       0:04:17.315 **** 
skipping: [managed-node3] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.027)       0:04:17.342 **** 
skipping: [managed-node3] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.026)       0:04:17.368 **** 
skipping: [managed-node3] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.025)       0:04:17.394 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.025)       0:04:17.419 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.022)       0:04:17.441 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.022)       0:04:17.464 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.022)       0:04:17.487 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.022)       0:04:17.509 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.020)       0:04:17.529 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.021)       0:04:17.551 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.019)       0:04:17.571 **** 
skipping: [managed-node3] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.021)       0:04:17.592 **** 
skipping: [managed-node3] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.024)       0:04:17.616 **** 
skipping: [managed-node3] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.020)       0:04:17.637 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.020)       0:04:17.658 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.021)       0:04:17.679 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.021)       0:04:17.700 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.021)       0:04:17.722 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.023)       0:04:17.746 **** 
ok: [managed-node3] => {
    "storage_test_actual_size": {
        "bytes": 5368709120,
        "changed": false,
        "failed": false,
        "lvm": "5g",
        "parted": "5GiB",
        "size": "5 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.024)       0:04:17.770 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.023)       0:04:17.794 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.031)       0:04:17.826 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.025972",
    "end": "2024-12-11 10:11:43.880189",
    "rc": 0,
    "start": "2024-12-11 10:11:43.854217"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.409)       0:04:18.235 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.028)       0:04:18.264 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Wednesday 11 December 2024  10:11:43 -0500 (0:00:00.028)       0:04:18.293 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.075)       0:04:18.369 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.036)       0:04:18.405 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.029)       0:04:18.435 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.027)       0:04:18.462 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.025)       0:04:18.487 **** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.023)       0:04:18.511 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Change volume size to 9g] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:268
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.029)       0:04:18.540 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.086)       0:04:18.627 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.031)       0:04:18.658 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.025)       0:04:18.684 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.059)       0:04:18.743 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.024)       0:04:18.768 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.025)       0:04:18.793 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.032)       0:04:18.826 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.034)       0:04:18.861 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.077)       0:04:18.938 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.033)       0:04:18.972 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext2",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "9g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.038)       0:04:19.011 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.027)       0:04:19.039 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.023)       0:04:19.062 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.028)       0:04:19.091 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.028)       0:04:19.119 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.031)       0:04:19.150 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.041)       0:04:19.192 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:11:44 -0500 (0:00:00.019)       0:04:19.211 **** 
changed: [managed-node3] => {
    "actions": [
        {
            "action": "resize device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "resize format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext2"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/mapper/foo-test1",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext2",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs",
        "lvm2"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext2",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "9g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:11:49 -0500 (0:00:04.578)       0:04:23.789 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:11:49 -0500 (0:00:00.028)       0:04:23.818 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929887.8428855,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "b54b2bd0584a8d1832fcc765abdde177fea1c71b",
        "ctime": 1733929887.8398855,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929887.8398855,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1394,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:11:49 -0500 (0:00:00.401)       0:04:24.219 **** 
ok: [managed-node3] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:11:50 -0500 (0:00:00.450)       0:04:24.670 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:11:50 -0500 (0:00:00.031)       0:04:24.701 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [
            {
                "action": "resize device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "resize format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext2"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/mapper/foo-test1",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext2",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs",
            "lvm2"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext2",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "9g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:11:50 -0500 (0:00:00.042)       0:04:24.743 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext2",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "9g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:11:50 -0500 (0:00:00.037)       0:04:24.781 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:11:50 -0500 (0:00:00.036)       0:04:24.818 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:11:50 -0500 (0:00:00.031)       0:04:24.849 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:11:51 -0500 (0:00:00.749)       0:04:25.599 **** 
ok: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext2', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext2",
    "mount_info": {
        "dump": 0,
        "fstype": "ext2",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:11:51 -0500 (0:00:00.427)       0:04:26.026 **** 
skipping: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext2', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "ext2",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:11:51 -0500 (0:00:00.046)       0:04:26.073 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:11:52 -0500 (0:00:00.683)       0:04:26.757 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:11:52 -0500 (0:00:00.438)       0:04:27.195 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:11:52 -0500 (0:00:00.021)       0:04:27.217 **** 
ok: [managed-node3]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:282
Wednesday 11 December 2024  10:11:53 -0500 (0:00:00.764)       0:04:27.981 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node3

TASK [Print out pool information] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Wednesday 11 December 2024  10:11:53 -0500 (0:00:00.055)       0:04:28.037 **** 
ok: [managed-node3] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext2",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "9g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Wednesday 11 December 2024  10:11:53 -0500 (0:00:00.030)       0:04:28.068 **** 
skipping: [managed-node3] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Wednesday 11 December 2024  10:11:53 -0500 (0:00:00.022)       0:04:28.090 **** 
ok: [managed-node3] => {
    "changed": false,
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "ext2",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/foo-test1",
            "size": "9G",
            "type": "lvm",
            "uuid": "78de8819-280f-4432-9837-123353bb821c"
        },
        "/dev/sda": {
            "fstype": "LVM2_member",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "VKmjYR-rX7d-wXez-nwPP-hiIt-jeKd-MyNwyV"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Wednesday 11 December 2024  10:11:54 -0500 (0:00:00.401)       0:04:28.492 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003395",
    "end": "2024-12-11 10:11:54.561775",
    "rc": 0,
    "start": "2024-12-11 10:11:54.558380"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/foo-test1 /opt/test1 ext2 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Wednesday 11 December 2024  10:11:54 -0500 (0:00:00.437)       0:04:28.929 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002660",
    "end": "2024-12-11 10:11:54.986181",
    "failed_when_result": false,
    "rc": 0,
    "start": "2024-12-11 10:11:54.983521"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Wednesday 11 December 2024  10:11:55 -0500 (0:00:00.428)       0:04:29.358 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node3

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Wednesday 11 December 2024  10:11:55 -0500 (0:00:00.070)       0:04:29.428 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Wednesday 11 December 2024  10:11:55 -0500 (0:00:00.049)       0:04:29.477 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "vgs",
        "--noheadings",
        "--binary",
        "-o",
        "shared",
        "foo"
    ],
    "delta": "0:00:00.030030",
    "end": "2024-12-11 10:11:55.564010",
    "rc": 0,
    "start": "2024-12-11 10:11:55.533980"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Wednesday 11 December 2024  10:11:55 -0500 (0:00:00.461)       0:04:29.938 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Wednesday 11 December 2024  10:11:55 -0500 (0:00:00.050)       0:04:29.989 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Wednesday 11 December 2024  10:11:55 -0500 (0:00:00.075)       0:04:30.064 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Wednesday 11 December 2024  10:11:55 -0500 (0:00:00.047)       0:04:30.112 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sda",
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Wednesday 11 December 2024  10:11:56 -0500 (0:00:00.431)       0:04:30.543 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Wednesday 11 December 2024  10:11:56 -0500 (0:00:00.095)       0:04:30.639 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Wednesday 11 December 2024  10:11:56 -0500 (0:00:00.091)       0:04:30.731 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Wednesday 11 December 2024  10:11:56 -0500 (0:00:00.045)       0:04:30.776 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Wednesday 11 December 2024  10:11:56 -0500 (0:00:00.040)       0:04:30.816 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Wednesday 11 December 2024  10:11:56 -0500 (0:00:00.053)       0:04:30.870 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Wednesday 11 December 2024  10:11:56 -0500 (0:00:00.047)       0:04:30.918 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Wednesday 11 December 2024  10:11:56 -0500 (0:00:00.060)       0:04:30.978 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.41.57 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.532)       0:04:31.511 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.042)       0:04:31.553 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node3

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.060)       0:04:31.613 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.025)       0:04:31.639 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.024)       0:04:31.664 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.027)       0:04:31.692 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.023)       0:04:31.715 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.028)       0:04:31.743 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.030)       0:04:31.773 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.026)       0:04:31.800 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.023)       0:04:31.824 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.022)       0:04:31.847 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.022)       0:04:31.869 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.021)       0:04:31.890 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node3

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.042)       0:04:31.933 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node3

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.044)       0:04:31.977 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.021)       0:04:31.999 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.020)       0:04:32.020 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.023)       0:04:32.043 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.020)       0:04:32.064 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.022)       0:04:32.086 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.027)       0:04:32.114 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.023)       0:04:32.137 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node3

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.068)       0:04:32.206 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node3

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Wednesday 11 December 2024  10:11:57 -0500 (0:00:00.057)       0:04:32.263 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.035)       0:04:32.299 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.031)       0:04:32.330 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.033)       0:04:32.364 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.032)       0:04:32.396 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.094)       0:04:32.490 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.055)       0:04:32.545 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.052)       0:04:32.598 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node3

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.054)       0:04:32.652 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.037)       0:04:32.690 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.030)       0:04:32.721 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.029)       0:04:32.750 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.024)       0:04:32.775 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.024)       0:04:32.799 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.022)       0:04:32.822 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.021)       0:04:32.843 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node3

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.062)       0:04:32.906 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node3

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.069)       0:04:32.976 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.035)       0:04:33.011 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.035)       0:04:33.047 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.027)       0:04:33.075 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.025)       0:04:33.101 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.026)       0:04:33.127 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.026)       0:04:33.154 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.028)       0:04:33.183 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node3

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Wednesday 11 December 2024  10:11:58 -0500 (0:00:00.101)       0:04:33.285 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.023)       0:04:33.308 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.029)       0:04:33.338 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.026)       0:04:33.364 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.037)       0:04:33.402 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.040)       0:04:33.443 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.038)       0:04:33.481 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.033)       0:04:33.514 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node3

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.061)       0:04:33.575 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.052)       0:04:33.628 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node3

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.168)       0:04:33.796 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.039)       0:04:33.835 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.042)       0:04:33.878 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.029)       0:04:33.907 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.036)       0:04:33.944 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.024)       0:04:33.968 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.024)       0:04:33.993 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.031)       0:04:34.024 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.023)       0:04:34.047 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.022)       0:04:34.070 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.024)       0:04:34.094 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.020)       0:04:34.115 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext2 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.046)       0:04:34.162 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.029)       0:04:34.192 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.038)       0:04:34.231 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Wednesday 11 December 2024  10:11:59 -0500 (0:00:00.034)       0:04:34.265 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Wednesday 11 December 2024  10:12:00 -0500 (0:00:00.040)       0:04:34.306 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Wednesday 11 December 2024  10:12:00 -0500 (0:00:00.037)       0:04:34.343 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Wednesday 11 December 2024  10:12:00 -0500 (0:00:00.047)       0:04:34.390 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Wednesday 11 December 2024  10:12:00 -0500 (0:00:00.054)       0:04:34.445 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929909.355815,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1733929909.355815,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 347287,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1733929909.355815,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Wednesday 11 December 2024  10:12:00 -0500 (0:00:00.412)       0:04:34.857 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Wednesday 11 December 2024  10:12:00 -0500 (0:00:00.027)       0:04:34.885 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Wednesday 11 December 2024  10:12:00 -0500 (0:00:00.022)       0:04:34.908 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Wednesday 11 December 2024  10:12:00 -0500 (0:00:00.026)       0:04:34.934 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Wednesday 11 December 2024  10:12:00 -0500 (0:00:00.024)       0:04:34.958 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Wednesday 11 December 2024  10:12:00 -0500 (0:00:00.021)       0:04:34.980 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Wednesday 11 December 2024  10:12:00 -0500 (0:00:00.024)       0:04:35.005 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Wednesday 11 December 2024  10:12:00 -0500 (0:00:00.020)       0:04:35.025 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Wednesday 11 December 2024  10:12:03 -0500 (0:00:02.838)       0:04:37.863 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Wednesday 11 December 2024  10:12:03 -0500 (0:00:00.023)       0:04:37.886 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Wednesday 11 December 2024  10:12:03 -0500 (0:00:00.023)       0:04:37.909 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Wednesday 11 December 2024  10:12:03 -0500 (0:00:00.031)       0:04:37.941 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Wednesday 11 December 2024  10:12:03 -0500 (0:00:00.022)       0:04:37.964 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Wednesday 11 December 2024  10:12:03 -0500 (0:00:00.023)       0:04:37.987 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Wednesday 11 December 2024  10:12:03 -0500 (0:00:00.023)       0:04:38.011 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Wednesday 11 December 2024  10:12:03 -0500 (0:00:00.066)       0:04:38.078 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Wednesday 11 December 2024  10:12:03 -0500 (0:00:00.022)       0:04:38.100 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Wednesday 11 December 2024  10:12:03 -0500 (0:00:00.030)       0:04:38.131 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Wednesday 11 December 2024  10:12:03 -0500 (0:00:00.031)       0:04:38.163 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Wednesday 11 December 2024  10:12:03 -0500 (0:00:00.027)       0:04:38.190 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Wednesday 11 December 2024  10:12:03 -0500 (0:00:00.033)       0:04:38.224 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Wednesday 11 December 2024  10:12:03 -0500 (0:00:00.039)       0:04:38.263 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Wednesday 11 December 2024  10:12:03 -0500 (0:00:00.031)       0:04:38.295 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Wednesday 11 December 2024  10:12:04 -0500 (0:00:00.033)       0:04:38.328 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Wednesday 11 December 2024  10:12:04 -0500 (0:00:00.035)       0:04:38.363 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Wednesday 11 December 2024  10:12:04 -0500 (0:00:00.036)       0:04:38.400 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Wednesday 11 December 2024  10:12:04 -0500 (0:00:00.023)       0:04:38.423 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Wednesday 11 December 2024  10:12:04 -0500 (0:00:00.027)       0:04:38.451 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Wednesday 11 December 2024  10:12:04 -0500 (0:00:00.028)       0:04:38.479 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Wednesday 11 December 2024  10:12:04 -0500 (0:00:00.026)       0:04:38.506 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Wednesday 11 December 2024  10:12:04 -0500 (0:00:00.026)       0:04:38.533 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Wednesday 11 December 2024  10:12:04 -0500 (0:00:00.026)       0:04:38.560 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Wednesday 11 December 2024  10:12:04 -0500 (0:00:00.022)       0:04:38.582 **** 
ok: [managed-node3] => {
    "bytes": 9663676416,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Wednesday 11 December 2024  10:12:04 -0500 (0:00:00.379)       0:04:38.962 **** 
ok: [managed-node3] => {
    "bytes": 9663676416,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.377)       0:04:39.340 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_expected_size": "9663676416"
    },
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.030)       0:04:39.371 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "9663676416"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.025)       0:04:39.396 **** 
ok: [managed-node3] => {
    "bytes": 10726680821,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.372)       0:04:39.769 **** 
skipping: [managed-node3] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.026)       0:04:39.795 **** 
skipping: [managed-node3] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.036)       0:04:39.832 **** 
skipping: [managed-node3] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.026)       0:04:39.859 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.026)       0:04:39.885 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.022)       0:04:39.907 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.021)       0:04:39.929 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.021)       0:04:39.950 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.021)       0:04:39.972 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.023)       0:04:39.995 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.022)       0:04:40.018 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.021)       0:04:40.040 **** 
skipping: [managed-node3] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.021)       0:04:40.061 **** 
skipping: [managed-node3] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.021)       0:04:40.083 **** 
skipping: [managed-node3] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.022)       0:04:40.105 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.024)       0:04:40.129 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.022)       0:04:40.152 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.022)       0:04:40.174 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.021)       0:04:40.195 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.021)       0:04:40.217 **** 
ok: [managed-node3] => {
    "storage_test_actual_size": {
        "bytes": 9663676416,
        "changed": false,
        "failed": false,
        "lvm": "9g",
        "parted": "9GiB",
        "size": "9 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.025)       0:04:40.242 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "9663676416"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Wednesday 11 December 2024  10:12:05 -0500 (0:00:00.027)       0:04:40.270 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Wednesday 11 December 2024  10:12:06 -0500 (0:00:00.033)       0:04:40.303 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.028691",
    "end": "2024-12-11 10:12:06.352055",
    "rc": 0,
    "start": "2024-12-11 10:12:06.323364"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Wednesday 11 December 2024  10:12:06 -0500 (0:00:00.404)       0:04:40.708 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Wednesday 11 December 2024  10:12:06 -0500 (0:00:00.029)       0:04:40.737 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Wednesday 11 December 2024  10:12:06 -0500 (0:00:00.030)       0:04:40.768 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Wednesday 11 December 2024  10:12:06 -0500 (0:00:00.025)       0:04:40.794 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Wednesday 11 December 2024  10:12:06 -0500 (0:00:00.028)       0:04:40.822 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Wednesday 11 December 2024  10:12:06 -0500 (0:00:00.030)       0:04:40.852 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Wednesday 11 December 2024  10:12:06 -0500 (0:00:00.035)       0:04:40.888 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Wednesday 11 December 2024  10:12:06 -0500 (0:00:00.023)       0:04:40.912 **** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Wednesday 11 December 2024  10:12:06 -0500 (0:00:00.035)       0:04:40.947 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Change volume size to 5g] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:285
Wednesday 11 December 2024  10:12:06 -0500 (0:00:00.033)       0:04:40.981 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:12:06 -0500 (0:00:00.129)       0:04:41.110 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:12:06 -0500 (0:00:00.103)       0:04:41.214 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:12:06 -0500 (0:00:00.039)       0:04:41.254 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:12:07 -0500 (0:00:00.103)       0:04:41.357 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:12:07 -0500 (0:00:00.049)       0:04:41.407 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:12:07 -0500 (0:00:00.060)       0:04:41.467 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:12:07 -0500 (0:00:00.049)       0:04:41.517 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:12:07 -0500 (0:00:00.048)       0:04:41.566 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:12:07 -0500 (0:00:00.098)       0:04:41.664 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:12:07 -0500 (0:00:00.044)       0:04:41.708 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext2",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "5g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:12:07 -0500 (0:00:00.046)       0:04:41.755 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:12:07 -0500 (0:00:00.036)       0:04:41.791 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:12:07 -0500 (0:00:00.043)       0:04:41.835 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:12:07 -0500 (0:00:00.035)       0:04:41.871 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:12:07 -0500 (0:00:00.039)       0:04:41.910 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:12:07 -0500 (0:00:00.049)       0:04:41.960 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:12:07 -0500 (0:00:00.052)       0:04:42.013 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:12:07 -0500 (0:00:00.029)       0:04:42.042 **** 
changed: [managed-node3] => {
    "actions": [
        {
            "action": "resize format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext2"
        },
        {
            "action": "resize device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/mapper/foo-test1",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext2",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "lvm2",
        "e2fsprogs",
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext2",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:12:13 -0500 (0:00:05.914)       0:04:47.957 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:12:13 -0500 (0:00:00.033)       0:04:47.990 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929887.8428855,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "b54b2bd0584a8d1832fcc765abdde177fea1c71b",
        "ctime": 1733929887.8398855,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929887.8398855,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1394,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:12:14 -0500 (0:00:00.480)       0:04:48.471 **** 
ok: [managed-node3] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:12:14 -0500 (0:00:00.536)       0:04:49.008 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:12:14 -0500 (0:00:00.035)       0:04:49.043 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [
            {
                "action": "resize format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext2"
            },
            {
                "action": "resize device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/mapper/foo-test1",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext2",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "lvm2",
            "e2fsprogs",
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext2",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:12:14 -0500 (0:00:00.042)       0:04:49.085 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext2",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:12:14 -0500 (0:00:00.053)       0:04:49.139 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:12:14 -0500 (0:00:00.053)       0:04:49.193 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:12:14 -0500 (0:00:00.029)       0:04:49.222 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:12:15 -0500 (0:00:00.674)       0:04:49.897 **** 
changed: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext2', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext2",
    "mount_info": {
        "dump": 0,
        "fstype": "ext2",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:12:16 -0500 (0:00:00.445)       0:04:50.342 **** 
skipping: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext2', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "ext2",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:12:16 -0500 (0:00:00.063)       0:04:50.406 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:12:16 -0500 (0:00:00.744)       0:04:51.150 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:12:17 -0500 (0:00:00.432)       0:04:51.582 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:12:17 -0500 (0:00:00.034)       0:04:51.617 **** 
ok: [managed-node3]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:299
Wednesday 11 December 2024  10:12:18 -0500 (0:00:00.829)       0:04:52.447 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node3

TASK [Print out pool information] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Wednesday 11 December 2024  10:12:18 -0500 (0:00:00.127)       0:04:52.574 **** 
ok: [managed-node3] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext2",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Wednesday 11 December 2024  10:12:18 -0500 (0:00:00.189)       0:04:52.764 **** 
skipping: [managed-node3] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Wednesday 11 December 2024  10:12:18 -0500 (0:00:00.049)       0:04:52.813 **** 
ok: [managed-node3] => {
    "changed": false,
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "ext2",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/foo-test1",
            "size": "5G",
            "type": "lvm",
            "uuid": "78de8819-280f-4432-9837-123353bb821c"
        },
        "/dev/sda": {
            "fstype": "LVM2_member",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "VKmjYR-rX7d-wXez-nwPP-hiIt-jeKd-MyNwyV"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Wednesday 11 December 2024  10:12:18 -0500 (0:00:00.463)       0:04:53.277 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002706",
    "end": "2024-12-11 10:12:19.480712",
    "rc": 0,
    "start": "2024-12-11 10:12:19.478006"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/foo-test1 /opt/test1 ext2 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Wednesday 11 December 2024  10:12:19 -0500 (0:00:00.621)       0:04:53.898 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002688",
    "end": "2024-12-11 10:12:19.962480",
    "failed_when_result": false,
    "rc": 0,
    "start": "2024-12-11 10:12:19.959792"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Wednesday 11 December 2024  10:12:20 -0500 (0:00:00.433)       0:04:54.332 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node3

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Wednesday 11 December 2024  10:12:20 -0500 (0:00:00.069)       0:04:54.402 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Wednesday 11 December 2024  10:12:20 -0500 (0:00:00.043)       0:04:54.446 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "vgs",
        "--noheadings",
        "--binary",
        "-o",
        "shared",
        "foo"
    ],
    "delta": "0:00:00.021097",
    "end": "2024-12-11 10:12:20.520141",
    "rc": 0,
    "start": "2024-12-11 10:12:20.499044"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Wednesday 11 December 2024  10:12:20 -0500 (0:00:00.443)       0:04:54.889 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Wednesday 11 December 2024  10:12:20 -0500 (0:00:00.046)       0:04:54.936 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Wednesday 11 December 2024  10:12:20 -0500 (0:00:00.072)       0:04:55.008 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Wednesday 11 December 2024  10:12:20 -0500 (0:00:00.048)       0:04:55.057 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sda",
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Wednesday 11 December 2024  10:12:21 -0500 (0:00:00.440)       0:04:55.498 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Wednesday 11 December 2024  10:12:21 -0500 (0:00:00.035)       0:04:55.533 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Wednesday 11 December 2024  10:12:21 -0500 (0:00:00.038)       0:04:55.571 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Wednesday 11 December 2024  10:12:21 -0500 (0:00:00.042)       0:04:55.614 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Wednesday 11 December 2024  10:12:21 -0500 (0:00:00.037)       0:04:55.651 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Wednesday 11 December 2024  10:12:21 -0500 (0:00:00.041)       0:04:55.693 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Wednesday 11 December 2024  10:12:21 -0500 (0:00:00.035)       0:04:55.729 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Wednesday 11 December 2024  10:12:21 -0500 (0:00:00.058)       0:04:55.787 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.41.57 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Wednesday 11 December 2024  10:12:21 -0500 (0:00:00.477)       0:04:56.265 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Wednesday 11 December 2024  10:12:21 -0500 (0:00:00.025)       0:04:56.291 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node3

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.041)       0:04:56.332 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.022)       0:04:56.355 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.021)       0:04:56.377 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.020)       0:04:56.398 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.027)       0:04:56.425 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.024)       0:04:56.449 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.040)       0:04:56.489 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.046)       0:04:56.535 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.048)       0:04:56.584 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.047)       0:04:56.632 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.035)       0:04:56.667 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.032)       0:04:56.700 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node3

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.074)       0:04:56.774 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node3

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.125)       0:04:56.900 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.057)       0:04:56.958 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.024)       0:04:56.982 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.101)       0:04:57.084 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.035)       0:04:57.120 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.054)       0:04:57.175 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.038)       0:04:57.214 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Wednesday 11 December 2024  10:12:22 -0500 (0:00:00.036)       0:04:57.250 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node3

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.061)       0:04:57.311 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node3

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.048)       0:04:57.360 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.022)       0:04:57.382 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.022)       0:04:57.404 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.023)       0:04:57.428 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.020)       0:04:57.448 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.047)       0:04:57.496 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.026)       0:04:57.523 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.028)       0:04:57.552 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node3

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.050)       0:04:57.603 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.026)       0:04:57.630 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.027)       0:04:57.657 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.027)       0:04:57.685 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.024)       0:04:57.709 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.032)       0:04:57.742 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.034)       0:04:57.776 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.038)       0:04:57.815 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node3

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.076)       0:04:57.891 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node3

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.067)       0:04:57.959 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.034)       0:04:57.993 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.033)       0:04:58.027 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.026)       0:04:58.053 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.024)       0:04:58.078 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.031)       0:04:58.109 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.023)       0:04:58.132 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.030)       0:04:58.162 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node3

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.056)       0:04:58.219 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.021)       0:04:58.241 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.021)       0:04:58.263 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Wednesday 11 December 2024  10:12:23 -0500 (0:00:00.021)       0:04:58.284 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.024)       0:04:58.309 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.021)       0:04:58.331 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.020)       0:04:58.352 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.020)       0:04:58.372 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node3

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.039)       0:04:58.412 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.029)       0:04:58.442 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node3

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.113)       0:04:58.555 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.026)       0:04:58.582 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.035)       0:04:58.618 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.080)       0:04:58.699 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.027)       0:04:58.726 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.023)       0:04:58.749 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.023)       0:04:58.773 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.023)       0:04:58.796 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.023)       0:04:58.819 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.025)       0:04:58.844 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.022)       0:04:58.867 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.022)       0:04:58.890 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext2 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.045)       0:04:58.936 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.028)       0:04:58.965 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.026)       0:04:58.991 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.025)       0:04:59.017 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.028)       0:04:59.045 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.023)       0:04:59.068 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.031)       0:04:59.100 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Wednesday 11 December 2024  10:12:24 -0500 (0:00:00.033)       0:04:59.134 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929933.491736,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1733929933.491736,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 347287,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1733929933.491736,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Wednesday 11 December 2024  10:12:25 -0500 (0:00:00.382)       0:04:59.516 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Wednesday 11 December 2024  10:12:25 -0500 (0:00:00.029)       0:04:59.546 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Wednesday 11 December 2024  10:12:25 -0500 (0:00:00.023)       0:04:59.569 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Wednesday 11 December 2024  10:12:25 -0500 (0:00:00.027)       0:04:59.597 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Wednesday 11 December 2024  10:12:25 -0500 (0:00:00.024)       0:04:59.621 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Wednesday 11 December 2024  10:12:25 -0500 (0:00:00.031)       0:04:59.653 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Wednesday 11 December 2024  10:12:25 -0500 (0:00:00.032)       0:04:59.685 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Wednesday 11 December 2024  10:12:25 -0500 (0:00:00.024)       0:04:59.710 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Wednesday 11 December 2024  10:12:28 -0500 (0:00:02.810)       0:05:02.521 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.028)       0:05:02.550 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.023)       0:05:02.573 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.047)       0:05:02.620 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.034)       0:05:02.655 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.035)       0:05:02.690 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.049)       0:05:02.739 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.042)       0:05:02.782 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.054)       0:05:02.836 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.066)       0:05:02.903 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.048)       0:05:02.951 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.035)       0:05:02.987 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.035)       0:05:03.022 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.033)       0:05:03.056 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.035)       0:05:03.092 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.032)       0:05:03.124 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.035)       0:05:03.160 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.032)       0:05:03.192 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.035)       0:05:03.227 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.032)       0:05:03.259 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Wednesday 11 December 2024  10:12:28 -0500 (0:00:00.036)       0:05:03.296 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Wednesday 11 December 2024  10:12:29 -0500 (0:00:00.031)       0:05:03.328 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Wednesday 11 December 2024  10:12:29 -0500 (0:00:00.033)       0:05:03.361 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Wednesday 11 December 2024  10:12:29 -0500 (0:00:00.048)       0:05:03.409 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Wednesday 11 December 2024  10:12:29 -0500 (0:00:00.066)       0:05:03.475 **** 
ok: [managed-node3] => {
    "bytes": 5368709120,
    "changed": false,
    "lvm": "5g",
    "parted": "5GiB",
    "size": "5 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Wednesday 11 December 2024  10:12:29 -0500 (0:00:00.521)       0:05:03.997 **** 
ok: [managed-node3] => {
    "bytes": 5368709120,
    "changed": false,
    "lvm": "5g",
    "parted": "5GiB",
    "size": "5 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.428)       0:05:04.426 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_expected_size": "5368709120"
    },
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.030)       0:05:04.456 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.023)       0:05:04.480 **** 
ok: [managed-node3] => {
    "bytes": 10726680821,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.399)       0:05:04.880 **** 
skipping: [managed-node3] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.026)       0:05:04.906 **** 
skipping: [managed-node3] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.026)       0:05:04.932 **** 
skipping: [managed-node3] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.025)       0:05:04.958 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.025)       0:05:04.984 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.022)       0:05:05.006 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.021)       0:05:05.028 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.021)       0:05:05.050 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.024)       0:05:05.074 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.066)       0:05:05.141 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.023)       0:05:05.164 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.021)       0:05:05.186 **** 
skipping: [managed-node3] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.021)       0:05:05.208 **** 
skipping: [managed-node3] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.022)       0:05:05.231 **** 
skipping: [managed-node3] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.021)       0:05:05.252 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Wednesday 11 December 2024  10:12:30 -0500 (0:00:00.024)       0:05:05.277 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.028)       0:05:05.305 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.026)       0:05:05.332 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.032)       0:05:05.365 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.034)       0:05:05.399 **** 
ok: [managed-node3] => {
    "storage_test_actual_size": {
        "bytes": 5368709120,
        "changed": false,
        "failed": false,
        "lvm": "5g",
        "parted": "5GiB",
        "size": "5 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.036)       0:05:05.436 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.039)       0:05:05.476 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.048)       0:05:05.524 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.026804",
    "end": "2024-12-11 10:12:31.581621",
    "rc": 0,
    "start": "2024-12-11 10:12:31.554817"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.414)       0:05:05.939 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.028)       0:05:05.968 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.030)       0:05:05.998 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.025)       0:05:06.023 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.026)       0:05:06.050 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.024)       0:05:06.074 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.024)       0:05:06.098 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.020)       0:05:06.119 **** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.018)       0:05:06.138 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Clean up] ****************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:302
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.020)       0:05:06.159 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.088)       0:05:06.247 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:12:31 -0500 (0:00:00.031)       0:05:06.279 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:12:32 -0500 (0:00:00.025)       0:05:06.304 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:12:32 -0500 (0:00:00.056)       0:05:06.361 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:12:32 -0500 (0:00:00.024)       0:05:06.385 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:12:32 -0500 (0:00:00.023)       0:05:06.409 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:12:32 -0500 (0:00:00.022)       0:05:06.431 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:12:32 -0500 (0:00:00.023)       0:05:06.455 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:12:32 -0500 (0:00:00.053)       0:05:06.508 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:12:32 -0500 (0:00:00.023)       0:05:06.532 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "state": "absent",
            "volumes": [
                {
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "5g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:12:32 -0500 (0:00:00.027)       0:05:06.560 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:12:32 -0500 (0:00:00.025)       0:05:06.585 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:12:32 -0500 (0:00:00.025)       0:05:06.611 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:12:32 -0500 (0:00:00.025)       0:05:06.636 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:12:32 -0500 (0:00:00.023)       0:05:06.660 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:12:32 -0500 (0:00:00.022)       0:05:06.683 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:12:32 -0500 (0:00:00.035)       0:05:06.719 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:12:32 -0500 (0:00:00.021)       0:05:06.740 **** 
changed: [managed-node3] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext2"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sda",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1"
    ],
    "mounts": [
        {
            "fstype": "ext2",
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "absent",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext2",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:12:37 -0500 (0:00:04.829)       0:05:11.569 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:12:37 -0500 (0:00:00.025)       0:05:11.595 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929887.8428855,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "b54b2bd0584a8d1832fcc765abdde177fea1c71b",
        "ctime": 1733929887.8398855,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929887.8398855,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1394,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:12:37 -0500 (0:00:00.390)       0:05:11.986 **** 
ok: [managed-node3] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:12:38 -0500 (0:00:00.386)       0:05:12.372 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:12:38 -0500 (0:00:00.019)       0:05:12.392 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext2"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sda",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1"
        ],
        "mounts": [
            {
                "fstype": "ext2",
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "absent",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext2",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:12:38 -0500 (0:00:00.027)       0:05:12.420 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "absent",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext2",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:12:38 -0500 (0:00:00.027)       0:05:12.448 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:12:38 -0500 (0:00:00.026)       0:05:12.474 **** 
changed: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'ext2'}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext2",
    "mount_info": {
        "fstype": "ext2",
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:12:38 -0500 (0:00:00.386)       0:05:12.861 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:12:39 -0500 (0:00:00.656)       0:05:13.517 **** 

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:12:39 -0500 (0:00:00.029)       0:05:13.547 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:12:39 -0500 (0:00:00.023)       0:05:13.570 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:12:39 -0500 (0:00:00.674)       0:05:14.244 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:12:40 -0500 (0:00:00.408)       0:05:14.653 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:12:40 -0500 (0:00:00.037)       0:05:14.690 **** 
ok: [managed-node3]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:315
Wednesday 11 December 2024  10:12:41 -0500 (0:00:00.807)       0:05:15.497 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node3

TASK [Print out pool information] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Wednesday 11 December 2024  10:12:41 -0500 (0:00:00.096)       0:05:15.594 **** 
ok: [managed-node3] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "absent",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext2",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Wednesday 11 December 2024  10:12:41 -0500 (0:00:00.045)       0:05:15.639 **** 
skipping: [managed-node3] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Wednesday 11 December 2024  10:12:41 -0500 (0:00:00.037)       0:05:15.677 **** 
ok: [managed-node3] => {
    "changed": false,
    "info": {
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Wednesday 11 December 2024  10:12:41 -0500 (0:00:00.430)       0:05:16.107 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003113",
    "end": "2024-12-11 10:12:42.197392",
    "rc": 0,
    "start": "2024-12-11 10:12:42.194279"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Wednesday 11 December 2024  10:12:42 -0500 (0:00:00.463)       0:05:16.570 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002682",
    "end": "2024-12-11 10:12:42.635731",
    "failed_when_result": false,
    "rc": 0,
    "start": "2024-12-11 10:12:42.633049"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Wednesday 11 December 2024  10:12:42 -0500 (0:00:00.433)       0:05:17.004 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node3

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Wednesday 11 December 2024  10:12:42 -0500 (0:00:00.069)       0:05:17.073 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Wednesday 11 December 2024  10:12:42 -0500 (0:00:00.029)       0:05:17.103 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Wednesday 11 December 2024  10:12:42 -0500 (0:00:00.032)       0:05:17.136 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Wednesday 11 December 2024  10:12:42 -0500 (0:00:00.034)       0:05:17.171 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Wednesday 11 December 2024  10:12:43 -0500 (0:00:00.140)       0:05:17.311 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "0",
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Wednesday 11 December 2024  10:12:43 -0500 (0:00:00.047)       0:05:17.359 **** 

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Wednesday 11 December 2024  10:12:43 -0500 (0:00:00.037)       0:05:17.397 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": "0"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Wednesday 11 December 2024  10:12:43 -0500 (0:00:00.043)       0:05:17.440 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": []
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Wednesday 11 December 2024  10:12:43 -0500 (0:00:00.064)       0:05:17.505 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Wednesday 11 December 2024  10:12:43 -0500 (0:00:00.076)       0:05:17.581 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Wednesday 11 December 2024  10:12:43 -0500 (0:00:00.084)       0:05:17.666 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Wednesday 11 December 2024  10:12:43 -0500 (0:00:00.041)       0:05:17.707 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Wednesday 11 December 2024  10:12:43 -0500 (0:00:00.052)       0:05:17.760 **** 

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Wednesday 11 December 2024  10:12:43 -0500 (0:00:00.050)       0:05:17.810 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.41.57 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.493)       0:05:18.304 **** 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.020)       0:05:18.325 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node3

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.043)       0:05:18.368 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.022)       0:05:18.390 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.024)       0:05:18.415 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.021)       0:05:18.436 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.021)       0:05:18.457 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.023)       0:05:18.481 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.026)       0:05:18.508 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.031)       0:05:18.539 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.035)       0:05:18.575 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.032)       0:05:18.608 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.032)       0:05:18.641 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.035)       0:05:18.676 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node3

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.066)       0:05:18.743 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node3

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.067)       0:05:18.811 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.038)       0:05:18.849 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.035)       0:05:18.884 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.033)       0:05:18.917 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.022)       0:05:18.940 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.025)       0:05:18.965 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.024)       0:05:18.989 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.026)       0:05:19.016 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node3

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.053)       0:05:19.069 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node3

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.047)       0:05:19.117 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.023)       0:05:19.140 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.020)       0:05:19.161 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.022)       0:05:19.183 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.020)       0:05:19.204 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.047)       0:05:19.252 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Wednesday 11 December 2024  10:12:44 -0500 (0:00:00.025)       0:05:19.277 **** 

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.021)       0:05:19.299 **** 

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.029)       0:05:19.328 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.021)       0:05:19.350 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node3

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.047)       0:05:19.397 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node3

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.043)       0:05:19.441 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.021)       0:05:19.463 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.022)       0:05:19.485 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.060)       0:05:19.546 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.022)       0:05:19.568 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.022)       0:05:19.590 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.022)       0:05:19.613 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.020)       0:05:19.633 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node3

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.052)       0:05:19.686 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.022)       0:05:19.708 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.021)       0:05:19.730 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.022)       0:05:19.752 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.024)       0:05:19.776 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.022)       0:05:19.798 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.022)       0:05:19.821 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.021)       0:05:19.842 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node3

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.041)       0:05:19.883 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": false,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.029)       0:05:19.913 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node3

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.111)       0:05:20.024 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.025)       0:05:20.050 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.028)       0:05:20.078 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.024)       0:05:20.103 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.018)       0:05:20.122 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.022)       0:05:20.145 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.022)       0:05:20.167 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.021)       0:05:20.188 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.021)       0:05:20.210 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.022)       0:05:20.233 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.022)       0:05:20.255 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Wednesday 11 December 2024  10:12:45 -0500 (0:00:00.020)       0:05:20.276 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "0",
        "storage_test_fstab_expected_mount_options_matches": "0",
        "storage_test_fstab_expected_mount_point_matches": "0",
        "storage_test_fstab_id_matches": [],
        "storage_test_fstab_mount_options_matches": [],
        "storage_test_fstab_mount_point_matches": []
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Wednesday 11 December 2024  10:12:46 -0500 (0:00:00.044)       0:05:20.320 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Wednesday 11 December 2024  10:12:46 -0500 (0:00:00.029)       0:05:20.349 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Wednesday 11 December 2024  10:12:46 -0500 (0:00:00.029)       0:05:20.378 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Wednesday 11 December 2024  10:12:46 -0500 (0:00:00.022)       0:05:20.401 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Wednesday 11 December 2024  10:12:46 -0500 (0:00:00.026)       0:05:20.428 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Wednesday 11 December 2024  10:12:46 -0500 (0:00:00.021)       0:05:20.449 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Wednesday 11 December 2024  10:12:46 -0500 (0:00:00.022)       0:05:20.471 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Wednesday 11 December 2024  10:12:46 -0500 (0:00:00.022)       0:05:20.494 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "exists": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Wednesday 11 December 2024  10:12:46 -0500 (0:00:00.377)       0:05:20.872 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Wednesday 11 December 2024  10:12:46 -0500 (0:00:00.021)       0:05:20.893 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Wednesday 11 December 2024  10:12:46 -0500 (0:00:00.027)       0:05:20.920 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Wednesday 11 December 2024  10:12:46 -0500 (0:00:00.015)       0:05:20.936 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Wednesday 11 December 2024  10:12:46 -0500 (0:00:00.021)       0:05:20.958 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Wednesday 11 December 2024  10:12:46 -0500 (0:00:00.021)       0:05:20.979 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Wednesday 11 December 2024  10:12:46 -0500 (0:00:00.018)       0:05:20.997 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Wednesday 11 December 2024  10:12:46 -0500 (0:00:00.020)       0:05:21.017 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Wednesday 11 December 2024  10:12:49 -0500 (0:00:02.845)       0:05:23.863 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Wednesday 11 December 2024  10:12:49 -0500 (0:00:00.035)       0:05:23.899 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Wednesday 11 December 2024  10:12:49 -0500 (0:00:00.031)       0:05:23.930 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Wednesday 11 December 2024  10:12:49 -0500 (0:00:00.026)       0:05:23.956 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Wednesday 11 December 2024  10:12:49 -0500 (0:00:00.031)       0:05:23.987 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Wednesday 11 December 2024  10:12:49 -0500 (0:00:00.032)       0:05:24.020 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Wednesday 11 December 2024  10:12:49 -0500 (0:00:00.025)       0:05:24.045 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Wednesday 11 December 2024  10:12:49 -0500 (0:00:00.028)       0:05:24.073 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Wednesday 11 December 2024  10:12:49 -0500 (0:00:00.025)       0:05:24.099 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Wednesday 11 December 2024  10:12:49 -0500 (0:00:00.044)       0:05:24.143 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Wednesday 11 December 2024  10:12:49 -0500 (0:00:00.039)       0:05:24.183 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Wednesday 11 December 2024  10:12:49 -0500 (0:00:00.037)       0:05:24.220 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Wednesday 11 December 2024  10:12:49 -0500 (0:00:00.030)       0:05:24.250 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Wednesday 11 December 2024  10:12:49 -0500 (0:00:00.022)       0:05:24.272 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.024)       0:05:24.297 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.080)       0:05:24.378 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.024)       0:05:24.402 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.029)       0:05:24.432 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.030)       0:05:24.463 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.024)       0:05:24.488 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.022)       0:05:24.510 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.024)       0:05:24.534 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.022)       0:05:24.557 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.021)       0:05:24.578 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.021)       0:05:24.600 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.025)       0:05:24.626 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.023)       0:05:24.649 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.035)       0:05:24.685 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.036)       0:05:24.721 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.032)       0:05:24.754 **** 
skipping: [managed-node3] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.032)       0:05:24.786 **** 
skipping: [managed-node3] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.033)       0:05:24.820 **** 
skipping: [managed-node3] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.031)       0:05:24.851 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.037)       0:05:24.889 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.033)       0:05:24.922 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.034)       0:05:24.956 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.037)       0:05:24.994 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.033)       0:05:25.027 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.032)       0:05:25.060 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.035)       0:05:25.095 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.033)       0:05:25.129 **** 
skipping: [managed-node3] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.033)       0:05:25.163 **** 
skipping: [managed-node3] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.033)       0:05:25.196 **** 
skipping: [managed-node3] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.033)       0:05:25.230 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Wednesday 11 December 2024  10:12:50 -0500 (0:00:00.033)       0:05:25.263 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.035)       0:05:25.299 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.033)       0:05:25.332 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.033)       0:05:25.366 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.032)       0:05:25.398 **** 
ok: [managed-node3] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.036)       0:05:25.434 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.036)       0:05:25.471 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.036)       0:05:25.508 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.033)       0:05:25.541 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.030)       0:05:25.571 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.034)       0:05:25.606 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.032)       0:05:25.638 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.035)       0:05:25.673 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.038)       0:05:25.712 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.033)       0:05:25.746 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.032)       0:05:25.778 **** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.031)       0:05:25.810 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Gather package facts] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:318
Wednesday 11 December 2024  10:12:51 -0500 (0:00:00.030)       0:05:25.841 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "packages": {
            "NetworkManager": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "NetworkManager",
                    "release": "15.el8",
                    "source": "rpm",
                    "version": "1.40.16"
                }
            ],
            "NetworkManager-libnm": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "NetworkManager-libnm",
                    "release": "15.el8",
                    "source": "rpm",
                    "version": "1.40.16"
                }
            ],
            "NetworkManager-team": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "NetworkManager-team",
                    "release": "15.el8",
                    "source": "rpm",
                    "version": "1.40.16"
                }
            ],
            "NetworkManager-tui": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "NetworkManager-tui",
                    "release": "15.el8",
                    "source": "rpm",
                    "version": "1.40.16"
                }
            ],
            "acl": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "acl",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.2.53"
                }
            ],
            "aspell": [
                {
                    "arch": "x86_64",
                    "epoch": 12,
                    "name": "aspell",
                    "release": "22.el8",
                    "source": "rpm",
                    "version": "0.60.6.1"
                }
            ],
            "audit": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "audit",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "3.1.2"
                }
            ],
            "audit-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "audit-libs",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "3.1.2"
                }
            ],
            "authselect": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "authselect",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "1.2.6"
                }
            ],
            "authselect-compat": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "authselect-compat",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "1.2.6"
                }
            ],
            "authselect-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "authselect-libs",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "1.2.6"
                }
            ],
            "avahi-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "avahi-libs",
                    "release": "27.el8",
                    "source": "rpm",
                    "version": "0.7"
                }
            ],
            "basesystem": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "basesystem",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "11"
                }
            ],
            "bash": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "bash",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "4.4.20"
                }
            ],
            "bc": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "bc",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "1.07.1"
                }
            ],
            "beakerlib": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "beakerlib",
                    "release": "1.el8bkr",
                    "source": "rpm",
                    "version": "1.29.3"
                }
            ],
            "beakerlib-redhat": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "beakerlib-redhat",
                    "release": "33.el8bkr",
                    "source": "rpm",
                    "version": "1"
                }
            ],
            "bind-export-libs": [
                {
                    "arch": "x86_64",
                    "epoch": 32,
                    "name": "bind-export-libs",
                    "release": "13.el8",
                    "source": "rpm",
                    "version": "9.11.36"
                }
            ],
            "binutils": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "binutils",
                    "release": "123.el8",
                    "source": "rpm",
                    "version": "2.30"
                }
            ],
            "biosdevname": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "biosdevname",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.7.3"
                }
            ],
            "bison": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "bison",
                    "release": "10.el8",
                    "source": "rpm",
                    "version": "3.0.4"
                }
            ],
            "blivet-data": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "blivet-data",
                    "release": "8.el8",
                    "source": "rpm",
                    "version": "3.6.0"
                }
            ],
            "boost-atomic": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "boost-atomic",
                    "release": "13.el8",
                    "source": "rpm",
                    "version": "1.66.0"
                }
            ],
            "boost-chrono": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "boost-chrono",
                    "release": "13.el8",
                    "source": "rpm",
                    "version": "1.66.0"
                }
            ],
            "boost-date-time": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "boost-date-time",
                    "release": "13.el8",
                    "source": "rpm",
                    "version": "1.66.0"
                }
            ],
            "boost-filesystem": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "boost-filesystem",
                    "release": "13.el8",
                    "source": "rpm",
                    "version": "1.66.0"
                }
            ],
            "boost-system": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "boost-system",
                    "release": "13.el8",
                    "source": "rpm",
                    "version": "1.66.0"
                }
            ],
            "boost-thread": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "boost-thread",
                    "release": "13.el8",
                    "source": "rpm",
                    "version": "1.66.0"
                }
            ],
            "boost-timer": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "boost-timer",
                    "release": "13.el8",
                    "source": "rpm",
                    "version": "1.66.0"
                }
            ],
            "brotli": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "brotli",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "1.0.6"
                }
            ],
            "bzip2-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "bzip2-libs",
                    "release": "26.el8",
                    "source": "rpm",
                    "version": "1.0.6"
                }
            ],
            "c-ares": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "c-ares",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "1.13.0"
                }
            ],
            "ca-certificates": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "ca-certificates",
                    "release": "80.0.el8",
                    "source": "rpm",
                    "version": "2023.2.60_v7.0.306"
                }
            ],
            "centos-gpg-keys": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "centos-gpg-keys",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "8"
                }
            ],
            "centos-stream-release": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "centos-stream-release",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "8.6"
                }
            ],
            "centos-stream-repos": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "centos-stream-repos",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "8"
                }
            ],
            "checkpolicy": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "checkpolicy",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "2.9"
                }
            ],
            "chkconfig": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "chkconfig",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "1.19.2"
                }
            ],
            "chrony": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "chrony",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "4.5"
                }
            ],
            "clevis": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "clevis",
                    "release": "15.el8",
                    "source": "rpm",
                    "version": "15"
                }
            ],
            "clevis-luks": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "clevis-luks",
                    "release": "15.el8",
                    "source": "rpm",
                    "version": "15"
                }
            ],
            "cloud-init": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "cloud-init",
                    "release": "7.el8.2",
                    "source": "rpm",
                    "version": "23.4"
                }
            ],
            "cloud-utils-growpart": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "cloud-utils-growpart",
                    "release": "0.el8",
                    "source": "rpm",
                    "version": "0.33"
                }
            ],
            "coreutils": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "coreutils",
                    "release": "15.el8",
                    "source": "rpm",
                    "version": "8.30"
                }
            ],
            "coreutils-common": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "coreutils-common",
                    "release": "15.el8",
                    "source": "rpm",
                    "version": "8.30"
                }
            ],
            "cpio": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "cpio",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "2.12"
                }
            ],
            "cpp": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "cpp",
                    "release": "22.el8",
                    "source": "rpm",
                    "version": "8.5.0"
                }
            ],
            "cracklib": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "cracklib",
                    "release": "15.el8",
                    "source": "rpm",
                    "version": "2.9.6"
                }
            ],
            "cracklib-dicts": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "cracklib-dicts",
                    "release": "15.el8",
                    "source": "rpm",
                    "version": "2.9.6"
                }
            ],
            "cronie": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "cronie",
                    "release": "10.el8",
                    "source": "rpm",
                    "version": "1.5.2"
                }
            ],
            "cronie-anacron": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "cronie-anacron",
                    "release": "10.el8",
                    "source": "rpm",
                    "version": "1.5.2"
                }
            ],
            "crontabs": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "crontabs",
                    "release": "17.20190603git.el8",
                    "source": "rpm",
                    "version": "1.11"
                }
            ],
            "crypto-policies": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "crypto-policies",
                    "release": "1.git3177e06.el8",
                    "source": "rpm",
                    "version": "20230731"
                }
            ],
            "crypto-policies-scripts": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "crypto-policies-scripts",
                    "release": "1.git3177e06.el8",
                    "source": "rpm",
                    "version": "20230731"
                }
            ],
            "cryptsetup": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "cryptsetup",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "2.3.7"
                }
            ],
            "cryptsetup-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "cryptsetup-libs",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "2.3.7"
                }
            ],
            "curl": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "curl",
                    "release": "34.el8",
                    "source": "rpm",
                    "version": "7.61.1"
                }
            ],
            "cyrus-sasl-lib": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "cyrus-sasl-lib",
                    "release": "6.el8_5",
                    "source": "rpm",
                    "version": "2.1.27"
                }
            ],
            "daxctl-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "daxctl-libs",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "71.1"
                }
            ],
            "dbus": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "dbus",
                    "release": "26.el8",
                    "source": "rpm",
                    "version": "1.12.8"
                }
            ],
            "dbus-common": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "dbus-common",
                    "release": "26.el8",
                    "source": "rpm",
                    "version": "1.12.8"
                }
            ],
            "dbus-daemon": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "dbus-daemon",
                    "release": "26.el8",
                    "source": "rpm",
                    "version": "1.12.8"
                }
            ],
            "dbus-glib": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "dbus-glib",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.110"
                }
            ],
            "dbus-libs": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "dbus-libs",
                    "release": "26.el8",
                    "source": "rpm",
                    "version": "1.12.8"
                }
            ],
            "dbus-tools": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "dbus-tools",
                    "release": "26.el8",
                    "source": "rpm",
                    "version": "1.12.8"
                }
            ],
            "device-mapper": [
                {
                    "arch": "x86_64",
                    "epoch": 8,
                    "name": "device-mapper",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "1.02.181"
                }
            ],
            "device-mapper-event": [
                {
                    "arch": "x86_64",
                    "epoch": 8,
                    "name": "device-mapper-event",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "1.02.181"
                }
            ],
            "device-mapper-event-libs": [
                {
                    "arch": "x86_64",
                    "epoch": 8,
                    "name": "device-mapper-event-libs",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "1.02.181"
                }
            ],
            "device-mapper-libs": [
                {
                    "arch": "x86_64",
                    "epoch": 8,
                    "name": "device-mapper-libs",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "1.02.181"
                }
            ],
            "device-mapper-multipath": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "device-mapper-multipath",
                    "release": "41.el8",
                    "source": "rpm",
                    "version": "0.8.4"
                }
            ],
            "device-mapper-multipath-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "device-mapper-multipath-libs",
                    "release": "41.el8",
                    "source": "rpm",
                    "version": "0.8.4"
                }
            ],
            "device-mapper-persistent-data": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "device-mapper-persistent-data",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "0.9.0"
                }
            ],
            "dhcp-client": [
                {
                    "arch": "x86_64",
                    "epoch": 12,
                    "name": "dhcp-client",
                    "release": "50.el8",
                    "source": "rpm",
                    "version": "4.3.6"
                }
            ],
            "dhcp-common": [
                {
                    "arch": "noarch",
                    "epoch": 12,
                    "name": "dhcp-common",
                    "release": "50.el8",
                    "source": "rpm",
                    "version": "4.3.6"
                }
            ],
            "dhcp-libs": [
                {
                    "arch": "x86_64",
                    "epoch": 12,
                    "name": "dhcp-libs",
                    "release": "50.el8",
                    "source": "rpm",
                    "version": "4.3.6"
                }
            ],
            "diffutils": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "diffutils",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "3.6"
                }
            ],
            "dmidecode": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "dmidecode",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "3.3"
                }
            ],
            "dnf": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "dnf",
                    "release": "20.el8",
                    "source": "rpm",
                    "version": "4.7.0"
                }
            ],
            "dnf-data": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "dnf-data",
                    "release": "20.el8",
                    "source": "rpm",
                    "version": "4.7.0"
                }
            ],
            "dnf-plugins-core": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "dnf-plugins-core",
                    "release": "25.el8",
                    "source": "rpm",
                    "version": "4.0.21"
                }
            ],
            "dracut": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "dracut",
                    "release": "233.git20240115.el8",
                    "source": "rpm",
                    "version": "049"
                }
            ],
            "dracut-config-rescue": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "dracut-config-rescue",
                    "release": "233.git20240115.el8",
                    "source": "rpm",
                    "version": "049"
                }
            ],
            "dracut-network": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "dracut-network",
                    "release": "233.git20240115.el8",
                    "source": "rpm",
                    "version": "049"
                }
            ],
            "dracut-squash": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "dracut-squash",
                    "release": "233.git20240115.el8",
                    "source": "rpm",
                    "version": "049"
                }
            ],
            "dyninst": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "dyninst",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "12.1.0"
                }
            ],
            "e2fsprogs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "e2fsprogs",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "1.45.6"
                }
            ],
            "e2fsprogs-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "e2fsprogs-libs",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "1.45.6"
                }
            ],
            "efivar-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "efivar-libs",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "37"
                }
            ],
            "elfutils-debuginfod-client": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "elfutils-debuginfod-client",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.190"
                }
            ],
            "elfutils-debuginfod-client-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "elfutils-debuginfod-client-devel",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.190"
                }
            ],
            "elfutils-default-yama-scope": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "elfutils-default-yama-scope",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.190"
                }
            ],
            "elfutils-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "elfutils-devel",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.190"
                }
            ],
            "elfutils-libelf": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "elfutils-libelf",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.190"
                }
            ],
            "elfutils-libelf-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "elfutils-libelf-devel",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.190"
                }
            ],
            "elfutils-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "elfutils-libs",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.190"
                }
            ],
            "emacs-filesystem": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "emacs-filesystem",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "26.1"
                }
            ],
            "epel-release": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "epel-release",
                    "release": "19.el8",
                    "source": "rpm",
                    "version": "8"
                }
            ],
            "ethtool": [
                {
                    "arch": "x86_64",
                    "epoch": 2,
                    "name": "ethtool",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "5.13"
                }
            ],
            "expat": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "expat",
                    "release": "13.el8",
                    "source": "rpm",
                    "version": "2.2.5"
                }
            ],
            "file": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "file",
                    "release": "26.el8",
                    "source": "rpm",
                    "version": "5.33"
                }
            ],
            "file-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "file-libs",
                    "release": "26.el8",
                    "source": "rpm",
                    "version": "5.33"
                }
            ],
            "filesystem": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "filesystem",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "3.8"
                }
            ],
            "findutils": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "findutils",
                    "release": "22.el8",
                    "source": "rpm",
                    "version": "4.6.0"
                }
            ],
            "firewalld": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "firewalld",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "0.9.11"
                }
            ],
            "firewalld-filesystem": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "firewalld-filesystem",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "0.9.11"
                }
            ],
            "flex": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "flex",
                    "release": "9.el8",
                    "source": "rpm",
                    "version": "2.6.1"
                }
            ],
            "freetype": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "freetype",
                    "release": "9.el8",
                    "source": "rpm",
                    "version": "2.9.1"
                }
            ],
            "fuse-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "fuse-libs",
                    "release": "19.el8",
                    "source": "rpm",
                    "version": "2.9.7"
                }
            ],
            "gawk": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "gawk",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "4.2.1"
                }
            ],
            "gcc": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "gcc",
                    "release": "22.el8",
                    "source": "rpm",
                    "version": "8.5.0"
                }
            ],
            "gdbm": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "gdbm",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "1.18"
                }
            ],
            "gdbm-libs": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "gdbm-libs",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "1.18"
                }
            ],
            "gdisk": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "gdisk",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "1.0.3"
                }
            ],
            "geolite2-city": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "geolite2-city",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "20180605"
                }
            ],
            "geolite2-country": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "geolite2-country",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "20180605"
                }
            ],
            "gettext": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "gettext",
                    "release": "17.el8",
                    "source": "rpm",
                    "version": "0.19.8.1"
                }
            ],
            "gettext-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "gettext-libs",
                    "release": "17.el8",
                    "source": "rpm",
                    "version": "0.19.8.1"
                }
            ],
            "git": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "git",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "2.43.0"
                }
            ],
            "git-core": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "git-core",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "2.43.0"
                }
            ],
            "git-core-doc": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "git-core-doc",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "2.43.0"
                }
            ],
            "glib2": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "glib2",
                    "release": "163.el8",
                    "source": "rpm",
                    "version": "2.56.4"
                }
            ],
            "glibc": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "glibc",
                    "release": "251.el8.2",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "glibc-common": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "glibc-common",
                    "release": "251.el8.2",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "glibc-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "glibc-devel",
                    "release": "251.el8.2",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "glibc-gconv-extra": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "glibc-gconv-extra",
                    "release": "251.el8.2",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "glibc-headers": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "glibc-headers",
                    "release": "251.el8.2",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "glibc-langpack-en": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "glibc-langpack-en",
                    "release": "251.el8.2",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "gmp": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "gmp",
                    "release": "12.el8",
                    "source": "rpm",
                    "version": "6.1.2"
                }
            ],
            "gnupg2": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "gnupg2",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.2.20"
                }
            ],
            "gnupg2-smime": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "gnupg2-smime",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.2.20"
                }
            ],
            "gnutls": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "gnutls",
                    "release": "8.el8.1",
                    "source": "rpm",
                    "version": "3.6.16"
                }
            ],
            "gobject-introspection": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "gobject-introspection",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "1.56.1"
                }
            ],
            "gpg-pubkey": [
                {
                    "arch": null,
                    "epoch": null,
                    "name": "gpg-pubkey",
                    "release": "5ccc5b19",
                    "source": "rpm",
                    "version": "8483c65d"
                },
                {
                    "arch": null,
                    "epoch": null,
                    "name": "gpg-pubkey",
                    "release": "5cf7cefb",
                    "source": "rpm",
                    "version": "2f86d6a1"
                }
            ],
            "gpgme": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "gpgme",
                    "release": "12.el8",
                    "source": "rpm",
                    "version": "1.13.1"
                }
            ],
            "gpm-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "gpm-libs",
                    "release": "17.el8",
                    "source": "rpm",
                    "version": "1.20.7"
                }
            ],
            "grep": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "grep",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "3.1"
                }
            ],
            "groff-base": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "groff-base",
                    "release": "18.el8",
                    "source": "rpm",
                    "version": "1.22.3"
                }
            ],
            "grub2-common": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "grub2-common",
                    "release": "129.el8",
                    "source": "rpm",
                    "version": "2.02"
                }
            ],
            "grub2-pc": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "grub2-pc",
                    "release": "129.el8",
                    "source": "rpm",
                    "version": "2.02"
                }
            ],
            "grub2-pc-modules": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "grub2-pc-modules",
                    "release": "129.el8",
                    "source": "rpm",
                    "version": "2.02"
                }
            ],
            "grub2-tools": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "grub2-tools",
                    "release": "129.el8",
                    "source": "rpm",
                    "version": "2.02"
                }
            ],
            "grub2-tools-extra": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "grub2-tools-extra",
                    "release": "129.el8",
                    "source": "rpm",
                    "version": "2.02"
                }
            ],
            "grub2-tools-minimal": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "grub2-tools-minimal",
                    "release": "129.el8",
                    "source": "rpm",
                    "version": "2.02"
                }
            ],
            "grubby": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "grubby",
                    "release": "48.el8",
                    "source": "rpm",
                    "version": "8.40"
                }
            ],
            "gssproxy": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "gssproxy",
                    "release": "21.el8",
                    "source": "rpm",
                    "version": "0.8.0"
                }
            ],
            "gzip": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "gzip",
                    "release": "13.el8",
                    "source": "rpm",
                    "version": "1.9"
                }
            ],
            "hardlink": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "hardlink",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "1.3"
                }
            ],
            "hdparm": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "hdparm",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "9.54"
                }
            ],
            "hostname": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "hostname",
                    "release": "7.el8.0.1",
                    "source": "rpm",
                    "version": "3.20"
                }
            ],
            "hwdata": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "hwdata",
                    "release": "8.22.el8",
                    "source": "rpm",
                    "version": "0.314"
                }
            ],
            "ima-evm-utils": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "ima-evm-utils",
                    "release": "12.el8",
                    "source": "rpm",
                    "version": "1.3.2"
                }
            ],
            "info": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "info",
                    "release": "7.el8_5",
                    "source": "rpm",
                    "version": "6.5"
                }
            ],
            "initscripts": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "initscripts",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "10.00.18"
                }
            ],
            "ipcalc": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "ipcalc",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "0.2.4"
                }
            ],
            "iproute": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "iproute",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "6.2.0"
                }
            ],
            "iprutils": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "iprutils",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "2.4.19"
                }
            ],
            "ipset": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "ipset",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "7.1"
                }
            ],
            "ipset-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "ipset-libs",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "7.1"
                }
            ],
            "iptables": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "iptables",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "1.8.5"
                }
            ],
            "iptables-ebtables": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "iptables-ebtables",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "1.8.5"
                }
            ],
            "iptables-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "iptables-libs",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "1.8.5"
                }
            ],
            "iputils": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "iputils",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "20180629"
                }
            ],
            "irqbalance": [
                {
                    "arch": "x86_64",
                    "epoch": 2,
                    "name": "irqbalance",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "1.9.2"
                }
            ],
            "isl": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "isl",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "0.16.1"
                }
            ],
            "iwl100-firmware": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "iwl100-firmware",
                    "release": "121.el8.1",
                    "source": "rpm",
                    "version": "39.31.5.1"
                }
            ],
            "iwl1000-firmware": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "iwl1000-firmware",
                    "release": "121.el8.1",
                    "source": "rpm",
                    "version": "39.31.5.1"
                }
            ],
            "iwl105-firmware": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "iwl105-firmware",
                    "release": "121.el8.1",
                    "source": "rpm",
                    "version": "18.168.6.1"
                }
            ],
            "iwl135-firmware": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "iwl135-firmware",
                    "release": "121.el8.1",
                    "source": "rpm",
                    "version": "18.168.6.1"
                }
            ],
            "iwl2000-firmware": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "iwl2000-firmware",
                    "release": "121.el8.1",
                    "source": "rpm",
                    "version": "18.168.6.1"
                }
            ],
            "iwl2030-firmware": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "iwl2030-firmware",
                    "release": "121.el8.1",
                    "source": "rpm",
                    "version": "18.168.6.1"
                }
            ],
            "iwl3160-firmware": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "iwl3160-firmware",
                    "release": "121.el8.1",
                    "source": "rpm",
                    "version": "25.30.13.0"
                }
            ],
            "iwl5000-firmware": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "iwl5000-firmware",
                    "release": "121.el8.1",
                    "source": "rpm",
                    "version": "8.83.5.1_1"
                }
            ],
            "iwl5150-firmware": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "iwl5150-firmware",
                    "release": "121.el8.1",
                    "source": "rpm",
                    "version": "8.24.2.2"
                }
            ],
            "iwl6000-firmware": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "iwl6000-firmware",
                    "release": "121.el8.1",
                    "source": "rpm",
                    "version": "9.221.4.1"
                }
            ],
            "iwl6000g2a-firmware": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "iwl6000g2a-firmware",
                    "release": "121.el8.1",
                    "source": "rpm",
                    "version": "18.168.6.1"
                }
            ],
            "iwl6050-firmware": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "iwl6050-firmware",
                    "release": "121.el8.1",
                    "source": "rpm",
                    "version": "41.28.5.1"
                }
            ],
            "iwl7260-firmware": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "iwl7260-firmware",
                    "release": "121.el8.1",
                    "source": "rpm",
                    "version": "25.30.13.0"
                }
            ],
            "jansson": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "jansson",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "2.14"
                }
            ],
            "jose": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "jose",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "10"
                }
            ],
            "jq": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "jq",
                    "release": "9.el8",
                    "source": "rpm",
                    "version": "1.6"
                }
            ],
            "json-c": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "json-c",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "0.13.1"
                }
            ],
            "kbd": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "kbd",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "2.0.4"
                }
            ],
            "kbd-legacy": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "kbd-legacy",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "2.0.4"
                }
            ],
            "kbd-misc": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "kbd-misc",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "2.0.4"
                }
            ],
            "kernel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "kernel",
                    "release": "553.5.1.el8",
                    "source": "rpm",
                    "version": "4.18.0"
                }
            ],
            "kernel-core": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "kernel-core",
                    "release": "553.5.1.el8",
                    "source": "rpm",
                    "version": "4.18.0"
                }
            ],
            "kernel-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "kernel-devel",
                    "release": "553.5.1.el8",
                    "source": "rpm",
                    "version": "4.18.0"
                }
            ],
            "kernel-headers": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "kernel-headers",
                    "release": "553.5.1.el8",
                    "source": "rpm",
                    "version": "4.18.0"
                }
            ],
            "kernel-modules": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "kernel-modules",
                    "release": "553.5.1.el8",
                    "source": "rpm",
                    "version": "4.18.0"
                }
            ],
            "kernel-tools": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "kernel-tools",
                    "release": "553.5.1.el8",
                    "source": "rpm",
                    "version": "4.18.0"
                }
            ],
            "kernel-tools-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "kernel-tools-libs",
                    "release": "553.5.1.el8",
                    "source": "rpm",
                    "version": "4.18.0"
                }
            ],
            "kexec-tools": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "kexec-tools",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "2.0.26"
                }
            ],
            "keyutils": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "keyutils",
                    "release": "9.el8",
                    "source": "rpm",
                    "version": "1.5.10"
                }
            ],
            "keyutils-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "keyutils-libs",
                    "release": "9.el8",
                    "source": "rpm",
                    "version": "1.5.10"
                }
            ],
            "keyutils-libs-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "keyutils-libs-devel",
                    "release": "9.el8",
                    "source": "rpm",
                    "version": "1.5.10"
                }
            ],
            "kmod": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "kmod",
                    "release": "20.el8",
                    "source": "rpm",
                    "version": "25"
                }
            ],
            "kmod-kvdo": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "kmod-kvdo",
                    "release": "94.el8",
                    "source": "rpm",
                    "version": "6.2.8.7"
                }
            ],
            "kmod-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "kmod-libs",
                    "release": "20.el8",
                    "source": "rpm",
                    "version": "25"
                }
            ],
            "kpartx": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "kpartx",
                    "release": "41.el8",
                    "source": "rpm",
                    "version": "0.8.4"
                }
            ],
            "krb5-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "krb5-devel",
                    "release": "27.el8",
                    "source": "rpm",
                    "version": "1.18.2"
                }
            ],
            "krb5-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "krb5-libs",
                    "release": "27.el8",
                    "source": "rpm",
                    "version": "1.18.2"
                }
            ],
            "langpacks-en": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "langpacks-en",
                    "release": "12.el8",
                    "source": "rpm",
                    "version": "1.0"
                }
            ],
            "less": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "less",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "530"
                }
            ],
            "libacl": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libacl",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.2.53"
                }
            ],
            "libaio": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libaio",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "0.3.112"
                }
            ],
            "libarchive": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libarchive",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "3.3.3"
                }
            ],
            "libassuan": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libassuan",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.5.1"
                }
            ],
            "libattr": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libattr",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.4.48"
                }
            ],
            "libbasicobjects": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libbasicobjects",
                    "release": "40.el8",
                    "source": "rpm",
                    "version": "0.1.1"
                }
            ],
            "libblkid": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libblkid",
                    "release": "46.el8",
                    "source": "rpm",
                    "version": "2.32.1"
                }
            ],
            "libblockdev": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libblockdev",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "libblockdev-crypto": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libblockdev-crypto",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "libblockdev-dm": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libblockdev-dm",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "libblockdev-fs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libblockdev-fs",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "libblockdev-kbd": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libblockdev-kbd",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "libblockdev-loop": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libblockdev-loop",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "libblockdev-lvm": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libblockdev-lvm",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "libblockdev-mdraid": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libblockdev-mdraid",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "libblockdev-mpath": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libblockdev-mpath",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "libblockdev-nvdimm": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libblockdev-nvdimm",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "libblockdev-part": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libblockdev-part",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "libblockdev-swap": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libblockdev-swap",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "libblockdev-utils": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libblockdev-utils",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "libbpf": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libbpf",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "0.5.0"
                }
            ],
            "libbytesize": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libbytesize",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "1.4"
                }
            ],
            "libcap": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libcap",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "2.48"
                }
            ],
            "libcap-ng": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libcap-ng",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "0.7.11"
                }
            ],
            "libcollection": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libcollection",
                    "release": "40.el8",
                    "source": "rpm",
                    "version": "0.7.0"
                }
            ],
            "libcom_err": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libcom_err",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "1.45.6"
                }
            ],
            "libcom_err-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libcom_err-devel",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "1.45.6"
                }
            ],
            "libcomps": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libcomps",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "0.1.18"
                }
            ],
            "libcroco": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libcroco",
                    "release": "4.el8_2.1",
                    "source": "rpm",
                    "version": "0.6.12"
                }
            ],
            "libcurl": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libcurl",
                    "release": "34.el8",
                    "source": "rpm",
                    "version": "7.61.1"
                }
            ],
            "libdaemon": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libdaemon",
                    "release": "15.el8",
                    "source": "rpm",
                    "version": "0.14"
                }
            ],
            "libdb": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libdb",
                    "release": "42.el8_4",
                    "source": "rpm",
                    "version": "5.3.28"
                }
            ],
            "libdb-utils": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libdb-utils",
                    "release": "42.el8_4",
                    "source": "rpm",
                    "version": "5.3.28"
                }
            ],
            "libdhash": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libdhash",
                    "release": "40.el8",
                    "source": "rpm",
                    "version": "0.5.0"
                }
            ],
            "libdnf": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libdnf",
                    "release": "19.el8",
                    "source": "rpm",
                    "version": "0.63.0"
                }
            ],
            "libedit": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libedit",
                    "release": "23.20170329cvs.el8",
                    "source": "rpm",
                    "version": "3.1"
                }
            ],
            "libestr": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libestr",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "0.1.10"
                }
            ],
            "libev": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libev",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "4.24"
                }
            ],
            "libevent": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libevent",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "2.1.8"
                }
            ],
            "libfastjson": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libfastjson",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.99.9"
                }
            ],
            "libfdisk": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libfdisk",
                    "release": "46.el8",
                    "source": "rpm",
                    "version": "2.32.1"
                }
            ],
            "libffi": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libffi",
                    "release": "24.el8",
                    "source": "rpm",
                    "version": "3.1"
                }
            ],
            "libgcc": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libgcc",
                    "release": "22.el8",
                    "source": "rpm",
                    "version": "8.5.0"
                }
            ],
            "libgcrypt": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libgcrypt",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "1.8.5"
                }
            ],
            "libgomp": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libgomp",
                    "release": "22.el8",
                    "source": "rpm",
                    "version": "8.5.0"
                }
            ],
            "libgpg-error": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libgpg-error",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "1.31"
                }
            ],
            "libibverbs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libibverbs",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "48.0"
                }
            ],
            "libidn2": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libidn2",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "2.2.0"
                }
            ],
            "libini_config": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libini_config",
                    "release": "40.el8",
                    "source": "rpm",
                    "version": "1.3.1"
                }
            ],
            "libjose": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libjose",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "10"
                }
            ],
            "libkadm5": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libkadm5",
                    "release": "27.el8",
                    "source": "rpm",
                    "version": "1.18.2"
                }
            ],
            "libkcapi": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libkcapi",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "1.4.0"
                }
            ],
            "libkcapi-hmaccalc": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libkcapi-hmaccalc",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "1.4.0"
                }
            ],
            "libksba": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libksba",
                    "release": "9.el8",
                    "source": "rpm",
                    "version": "1.3.5"
                }
            ],
            "libldb": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libldb",
                    "release": "0.el8",
                    "source": "rpm",
                    "version": "2.8.0"
                }
            ],
            "libluksmeta": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libluksmeta",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "9"
                }
            ],
            "libmaxminddb": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libmaxminddb",
                    "release": "10.el8.1",
                    "source": "rpm",
                    "version": "1.2.0"
                }
            ],
            "libmetalink": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libmetalink",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "0.1.3"
                }
            ],
            "libmnl": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libmnl",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "1.0.4"
                }
            ],
            "libmodulemd": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libmodulemd",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "2.13.0"
                }
            ],
            "libmount": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libmount",
                    "release": "46.el8",
                    "source": "rpm",
                    "version": "2.32.1"
                }
            ],
            "libmpc": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libmpc",
                    "release": "9.1.el8",
                    "source": "rpm",
                    "version": "1.1.0"
                }
            ],
            "libndp": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libndp",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "1.7"
                }
            ],
            "libnetfilter_conntrack": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libnetfilter_conntrack",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "1.0.6"
                }
            ],
            "libnfnetlink": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libnfnetlink",
                    "release": "13.el8",
                    "source": "rpm",
                    "version": "1.0.1"
                }
            ],
            "libnfsidmap": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "libnfsidmap",
                    "release": "59.el8",
                    "source": "rpm",
                    "version": "2.3.3"
                }
            ],
            "libnftnl": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libnftnl",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "1.2.2"
                }
            ],
            "libnghttp2": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libnghttp2",
                    "release": "6.el8.1",
                    "source": "rpm",
                    "version": "1.33.0"
                }
            ],
            "libnl3": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libnl3",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "3.7.0"
                }
            ],
            "libnl3-cli": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libnl3-cli",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "3.7.0"
                }
            ],
            "libnsl2": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libnsl2",
                    "release": "2.20180605git4a062cf.el8",
                    "source": "rpm",
                    "version": "1.2.0"
                }
            ],
            "libpath_utils": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libpath_utils",
                    "release": "40.el8",
                    "source": "rpm",
                    "version": "0.2.1"
                }
            ],
            "libpcap": [
                {
                    "arch": "x86_64",
                    "epoch": 14,
                    "name": "libpcap",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "1.9.1"
                }
            ],
            "libpipeline": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libpipeline",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "1.5.0"
                }
            ],
            "libpkgconf": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libpkgconf",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "1.4.2"
                }
            ],
            "libpng": [
                {
                    "arch": "x86_64",
                    "epoch": 2,
                    "name": "libpng",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "1.6.34"
                }
            ],
            "libpsl": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libpsl",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "0.20.2"
                }
            ],
            "libpwquality": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libpwquality",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "1.4.4"
                }
            ],
            "libref_array": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libref_array",
                    "release": "40.el8",
                    "source": "rpm",
                    "version": "0.1.5"
                }
            ],
            "librepo": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "librepo",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "1.14.2"
                }
            ],
            "libreport-filesystem": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libreport-filesystem",
                    "release": "15.el8",
                    "source": "rpm",
                    "version": "2.9.5"
                }
            ],
            "libseccomp": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libseccomp",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "2.5.2"
                }
            ],
            "libsecret": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libsecret",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "0.18.6"
                }
            ],
            "libselinux": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libselinux",
                    "release": "8.el8",
                    "source": "rpm",
                    "version": "2.9"
                }
            ],
            "libselinux-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libselinux-devel",
                    "release": "8.el8",
                    "source": "rpm",
                    "version": "2.9"
                }
            ],
            "libselinux-utils": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libselinux-utils",
                    "release": "8.el8",
                    "source": "rpm",
                    "version": "2.9"
                }
            ],
            "libsemanage": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libsemanage",
                    "release": "9.el8",
                    "source": "rpm",
                    "version": "2.9"
                }
            ],
            "libsepol": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libsepol",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.9"
                }
            ],
            "libsepol-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libsepol-devel",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.9"
                }
            ],
            "libsigsegv": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libsigsegv",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "2.11"
                }
            ],
            "libsmartcols": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libsmartcols",
                    "release": "46.el8",
                    "source": "rpm",
                    "version": "2.32.1"
                }
            ],
            "libsolv": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libsolv",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "0.7.20"
                }
            ],
            "libss": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libss",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "1.45.6"
                }
            ],
            "libssh": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libssh",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "0.9.6"
                }
            ],
            "libssh-config": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "libssh-config",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "0.9.6"
                }
            ],
            "libsss_autofs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libsss_autofs",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.9.4"
                }
            ],
            "libsss_certmap": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libsss_certmap",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.9.4"
                }
            ],
            "libsss_idmap": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libsss_idmap",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.9.4"
                }
            ],
            "libsss_nss_idmap": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libsss_nss_idmap",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.9.4"
                }
            ],
            "libsss_sudo": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libsss_sudo",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.9.4"
                }
            ],
            "libstdc++": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libstdc++",
                    "release": "22.el8",
                    "source": "rpm",
                    "version": "8.5.0"
                }
            ],
            "libsysfs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libsysfs",
                    "release": "25.el8",
                    "source": "rpm",
                    "version": "2.1.0"
                }
            ],
            "libtalloc": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libtalloc",
                    "release": "0.el8",
                    "source": "rpm",
                    "version": "2.4.1"
                }
            ],
            "libtasn1": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libtasn1",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "4.13"
                }
            ],
            "libtdb": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libtdb",
                    "release": "0.el8",
                    "source": "rpm",
                    "version": "1.4.9"
                }
            ],
            "libteam": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libteam",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "1.31"
                }
            ],
            "libtevent": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libtevent",
                    "release": "0.el8",
                    "source": "rpm",
                    "version": "0.16.0"
                }
            ],
            "libtirpc": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libtirpc",
                    "release": "12.el8",
                    "source": "rpm",
                    "version": "1.1.4"
                }
            ],
            "libunistring": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libunistring",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "0.9.9"
                }
            ],
            "libusbx": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libusbx",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "1.0.23"
                }
            ],
            "libuser": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libuser",
                    "release": "25.el8",
                    "source": "rpm",
                    "version": "0.62"
                }
            ],
            "libutempter": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libutempter",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "1.1.6"
                }
            ],
            "libuuid": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libuuid",
                    "release": "46.el8",
                    "source": "rpm",
                    "version": "2.32.1"
                }
            ],
            "libverto": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libverto",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.3.2"
                }
            ],
            "libverto-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libverto-devel",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.3.2"
                }
            ],
            "libverto-libev": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libverto-libev",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.3.2"
                }
            ],
            "libxcrypt": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libxcrypt",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "4.1.1"
                }
            ],
            "libxcrypt-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libxcrypt-devel",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "4.1.1"
                }
            ],
            "libxkbcommon": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libxkbcommon",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "0.9.1"
                }
            ],
            "libxml2": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libxml2",
                    "release": "18.el8",
                    "source": "rpm",
                    "version": "2.9.7"
                }
            ],
            "libxslt": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libxslt",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "1.1.32"
                }
            ],
            "libyaml": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libyaml",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "0.1.7"
                }
            ],
            "libzstd": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libzstd",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "1.4.4"
                }
            ],
            "libzstd-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "libzstd-devel",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "1.4.4"
                }
            ],
            "linux-firmware": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "linux-firmware",
                    "release": "121.gitb3132c18.el8",
                    "source": "rpm",
                    "version": "20240111"
                }
            ],
            "lmdb-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "lmdb-libs",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.9.24"
                }
            ],
            "logrotate": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "logrotate",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "3.14.0"
                }
            ],
            "lshw": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "lshw",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "B.02.19.2"
                }
            ],
            "lsof": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "lsof",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "4.93.2"
                }
            ],
            "lsscsi": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "lsscsi",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "0.32"
                }
            ],
            "lua-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "lua-libs",
                    "release": "12.el8",
                    "source": "rpm",
                    "version": "5.3.4"
                }
            ],
            "luksmeta": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "luksmeta",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "9"
                }
            ],
            "lvm2": [
                {
                    "arch": "x86_64",
                    "epoch": 8,
                    "name": "lvm2",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "2.03.14"
                }
            ],
            "lvm2-libs": [
                {
                    "arch": "x86_64",
                    "epoch": 8,
                    "name": "lvm2-libs",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "2.03.14"
                }
            ],
            "lz4-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "lz4-libs",
                    "release": "3.el8_4",
                    "source": "rpm",
                    "version": "1.8.3"
                }
            ],
            "lzo": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "lzo",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "2.08"
                }
            ],
            "m4": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "m4",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "1.4.18"
                }
            ],
            "make": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "make",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "4.2.1"
                }
            ],
            "man-db": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "man-db",
                    "release": "18.el8",
                    "source": "rpm",
                    "version": "2.7.6.1"
                }
            ],
            "mdadm": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "mdadm",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "4.2"
                }
            ],
            "memstrack": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "memstrack",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.2.5"
                }
            ],
            "microcode_ctl": [
                {
                    "arch": "x86_64",
                    "epoch": 4,
                    "name": "microcode_ctl",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "20230808"
                }
            ],
            "mokutil": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "mokutil",
                    "release": "12.el8",
                    "source": "rpm",
                    "version": "0.3.0"
                }
            ],
            "mozjs60": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "mozjs60",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "60.9.0"
                }
            ],
            "mpfr": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "mpfr",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "3.1.6"
                }
            ],
            "ncurses": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "ncurses",
                    "release": "10.20180224.el8",
                    "source": "rpm",
                    "version": "6.1"
                }
            ],
            "ncurses-base": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "ncurses-base",
                    "release": "10.20180224.el8",
                    "source": "rpm",
                    "version": "6.1"
                }
            ],
            "ncurses-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "ncurses-libs",
                    "release": "10.20180224.el8",
                    "source": "rpm",
                    "version": "6.1"
                }
            ],
            "ndctl": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "ndctl",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "71.1"
                }
            ],
            "ndctl-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "ndctl-libs",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "71.1"
                }
            ],
            "nettle": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "nettle",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "3.4.1"
                }
            ],
            "newt": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "newt",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "0.52.20"
                }
            ],
            "nfs-utils": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "nfs-utils",
                    "release": "59.el8",
                    "source": "rpm",
                    "version": "2.3.3"
                }
            ],
            "nftables": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "nftables",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "1.0.4"
                }
            ],
            "npth": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "npth",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "1.5"
                }
            ],
            "nspr": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "nspr",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "4.35.0"
                }
            ],
            "nss": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "nss",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "3.90.0"
                }
            ],
            "nss-softokn": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "nss-softokn",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "3.90.0"
                }
            ],
            "nss-softokn-freebl": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "nss-softokn-freebl",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "3.90.0"
                }
            ],
            "nss-sysinit": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "nss-sysinit",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "3.90.0"
                }
            ],
            "nss-util": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "nss-util",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "3.90.0"
                }
            ],
            "numactl-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "numactl-libs",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "2.0.16"
                }
            ],
            "oddjob": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "oddjob",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "0.34.7"
                }
            ],
            "oddjob-mkhomedir": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "oddjob-mkhomedir",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "0.34.7"
                }
            ],
            "oniguruma": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "oniguruma",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "6.8.2"
                }
            ],
            "openldap": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "openldap",
                    "release": "18.el8",
                    "source": "rpm",
                    "version": "2.4.46"
                }
            ],
            "openssh": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "openssh",
                    "release": "24.el8",
                    "source": "rpm",
                    "version": "8.0p1"
                }
            ],
            "openssh-clients": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "openssh-clients",
                    "release": "24.el8",
                    "source": "rpm",
                    "version": "8.0p1"
                }
            ],
            "openssh-server": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "openssh-server",
                    "release": "24.el8",
                    "source": "rpm",
                    "version": "8.0p1"
                }
            ],
            "openssl": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "openssl",
                    "release": "12.el8",
                    "source": "rpm",
                    "version": "1.1.1k"
                }
            ],
            "openssl-devel": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "openssl-devel",
                    "release": "12.el8",
                    "source": "rpm",
                    "version": "1.1.1k"
                }
            ],
            "openssl-libs": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "openssl-libs",
                    "release": "12.el8",
                    "source": "rpm",
                    "version": "1.1.1k"
                }
            ],
            "openssl-pkcs11": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "openssl-pkcs11",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "0.4.10"
                }
            ],
            "os-prober": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "os-prober",
                    "release": "9.el8",
                    "source": "rpm",
                    "version": "1.74"
                }
            ],
            "p11-kit": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "p11-kit",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.23.22"
                }
            ],
            "p11-kit-trust": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "p11-kit-trust",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.23.22"
                }
            ],
            "pam": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "pam",
                    "release": "33.el8",
                    "source": "rpm",
                    "version": "1.3.1"
                }
            ],
            "parted": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "parted",
                    "release": "39.el8",
                    "source": "rpm",
                    "version": "3.2"
                }
            ],
            "passwd": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "passwd",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "0.80"
                }
            ],
            "pciutils-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "pciutils-libs",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "3.7.0"
                }
            ],
            "pcre": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "pcre",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "8.42"
                }
            ],
            "pcre2": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "pcre2",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "10.32"
                }
            ],
            "pcre2-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "pcre2-devel",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "10.32"
                }
            ],
            "pcre2-utf16": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "pcre2-utf16",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "10.32"
                }
            ],
            "pcre2-utf32": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "pcre2-utf32",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "10.32"
                }
            ],
            "perl-Carp": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-Carp",
                    "release": "396.el8",
                    "source": "rpm",
                    "version": "1.42"
                }
            ],
            "perl-Data-Dumper": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "perl-Data-Dumper",
                    "release": "399.el8",
                    "source": "rpm",
                    "version": "2.167"
                }
            ],
            "perl-Digest": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-Digest",
                    "release": "395.el8",
                    "source": "rpm",
                    "version": "1.17"
                }
            ],
            "perl-Digest-MD5": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "perl-Digest-MD5",
                    "release": "396.el8",
                    "source": "rpm",
                    "version": "2.55"
                }
            ],
            "perl-Encode": [
                {
                    "arch": "x86_64",
                    "epoch": 4,
                    "name": "perl-Encode",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.97"
                }
            ],
            "perl-Errno": [
                {
                    "arch": "x86_64",
                    "epoch": 0,
                    "name": "perl-Errno",
                    "release": "422.el8",
                    "source": "rpm",
                    "version": "1.28"
                }
            ],
            "perl-Error": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "perl-Error",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.17025"
                }
            ],
            "perl-Exporter": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-Exporter",
                    "release": "396.el8",
                    "source": "rpm",
                    "version": "5.72"
                }
            ],
            "perl-File-Path": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-File-Path",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "2.15"
                }
            ],
            "perl-File-Temp": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-File-Temp",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "0.230.600"
                }
            ],
            "perl-Getopt-Long": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "perl-Getopt-Long",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "2.50"
                }
            ],
            "perl-Git": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-Git",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "2.43.0"
                }
            ],
            "perl-HTTP-Tiny": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-HTTP-Tiny",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "0.074"
                }
            ],
            "perl-IO": [
                {
                    "arch": "x86_64",
                    "epoch": 0,
                    "name": "perl-IO",
                    "release": "422.el8",
                    "source": "rpm",
                    "version": "1.38"
                }
            ],
            "perl-IO-Socket-IP": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-IO-Socket-IP",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "0.39"
                }
            ],
            "perl-IO-Socket-SSL": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-IO-Socket-SSL",
                    "release": "4.module_el8+339+1ec643e0",
                    "source": "rpm",
                    "version": "2.066"
                }
            ],
            "perl-MIME-Base64": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "perl-MIME-Base64",
                    "release": "396.el8",
                    "source": "rpm",
                    "version": "3.15"
                }
            ],
            "perl-Mozilla-CA": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-Mozilla-CA",
                    "release": "7.module_el8+645+9d809f8c",
                    "source": "rpm",
                    "version": "20160104"
                }
            ],
            "perl-Net-SSLeay": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "perl-Net-SSLeay",
                    "release": "2.module_el8+339+1ec643e0",
                    "source": "rpm",
                    "version": "1.88"
                }
            ],
            "perl-PathTools": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "perl-PathTools",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "3.74"
                }
            ],
            "perl-Pod-Escapes": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "perl-Pod-Escapes",
                    "release": "395.el8",
                    "source": "rpm",
                    "version": "1.07"
                }
            ],
            "perl-Pod-Perldoc": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-Pod-Perldoc",
                    "release": "396.el8",
                    "source": "rpm",
                    "version": "3.28"
                }
            ],
            "perl-Pod-Simple": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "perl-Pod-Simple",
                    "release": "395.el8",
                    "source": "rpm",
                    "version": "3.35"
                }
            ],
            "perl-Pod-Usage": [
                {
                    "arch": "noarch",
                    "epoch": 4,
                    "name": "perl-Pod-Usage",
                    "release": "395.el8",
                    "source": "rpm",
                    "version": "1.69"
                }
            ],
            "perl-Scalar-List-Utils": [
                {
                    "arch": "x86_64",
                    "epoch": 3,
                    "name": "perl-Scalar-List-Utils",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "1.49"
                }
            ],
            "perl-Socket": [
                {
                    "arch": "x86_64",
                    "epoch": 4,
                    "name": "perl-Socket",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.027"
                }
            ],
            "perl-Storable": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "perl-Storable",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "3.11"
                }
            ],
            "perl-Term-ANSIColor": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-Term-ANSIColor",
                    "release": "396.el8",
                    "source": "rpm",
                    "version": "4.06"
                }
            ],
            "perl-Term-Cap": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-Term-Cap",
                    "release": "395.el8",
                    "source": "rpm",
                    "version": "1.17"
                }
            ],
            "perl-TermReadKey": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "perl-TermReadKey",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "2.37"
                }
            ],
            "perl-Text-ParseWords": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-Text-ParseWords",
                    "release": "395.el8",
                    "source": "rpm",
                    "version": "3.30"
                }
            ],
            "perl-Text-Tabs+Wrap": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-Text-Tabs+Wrap",
                    "release": "395.el8",
                    "source": "rpm",
                    "version": "2013.0523"
                }
            ],
            "perl-Time-Local": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "perl-Time-Local",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "1.280"
                }
            ],
            "perl-URI": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-URI",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "1.73"
                }
            ],
            "perl-Unicode-Normalize": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "perl-Unicode-Normalize",
                    "release": "396.el8",
                    "source": "rpm",
                    "version": "1.25"
                }
            ],
            "perl-constant": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-constant",
                    "release": "396.el8",
                    "source": "rpm",
                    "version": "1.33"
                }
            ],
            "perl-interpreter": [
                {
                    "arch": "x86_64",
                    "epoch": 4,
                    "name": "perl-interpreter",
                    "release": "422.el8",
                    "source": "rpm",
                    "version": "5.26.3"
                }
            ],
            "perl-libnet": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-libnet",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "3.11"
                }
            ],
            "perl-libs": [
                {
                    "arch": "x86_64",
                    "epoch": 4,
                    "name": "perl-libs",
                    "release": "422.el8",
                    "source": "rpm",
                    "version": "5.26.3"
                }
            ],
            "perl-macros": [
                {
                    "arch": "x86_64",
                    "epoch": 4,
                    "name": "perl-macros",
                    "release": "422.el8",
                    "source": "rpm",
                    "version": "5.26.3"
                }
            ],
            "perl-parent": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "perl-parent",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "0.237"
                }
            ],
            "perl-podlators": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "perl-podlators",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "4.11"
                }
            ],
            "perl-threads": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "perl-threads",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "2.21"
                }
            ],
            "perl-threads-shared": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "perl-threads-shared",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "1.58"
                }
            ],
            "pigz": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "pigz",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "2.4"
                }
            ],
            "pinentry": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "pinentry",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "1.1.0"
                }
            ],
            "pkgconf": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "pkgconf",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "1.4.2"
                }
            ],
            "pkgconf-m4": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "pkgconf-m4",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "1.4.2"
                }
            ],
            "pkgconf-pkg-config": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "pkgconf-pkg-config",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "1.4.2"
                }
            ],
            "platform-python": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "platform-python",
                    "release": "62.el8",
                    "source": "rpm",
                    "version": "3.6.8"
                }
            ],
            "platform-python-pip": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "platform-python-pip",
                    "release": "24.el8",
                    "source": "rpm",
                    "version": "9.0.3"
                }
            ],
            "platform-python-setuptools": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "platform-python-setuptools",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "39.2.0"
                }
            ],
            "plymouth": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "plymouth",
                    "release": "11.20200615git1e36e30.el8",
                    "source": "rpm",
                    "version": "0.9.4"
                }
            ],
            "plymouth-core-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "plymouth-core-libs",
                    "release": "11.20200615git1e36e30.el8",
                    "source": "rpm",
                    "version": "0.9.4"
                }
            ],
            "plymouth-scripts": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "plymouth-scripts",
                    "release": "11.20200615git1e36e30.el8",
                    "source": "rpm",
                    "version": "0.9.4"
                }
            ],
            "policycoreutils": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "policycoreutils",
                    "release": "26.el8",
                    "source": "rpm",
                    "version": "2.9"
                }
            ],
            "polkit": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "polkit",
                    "release": "15.el8",
                    "source": "rpm",
                    "version": "0.115"
                }
            ],
            "polkit-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "polkit-libs",
                    "release": "15.el8",
                    "source": "rpm",
                    "version": "0.115"
                }
            ],
            "polkit-pkla-compat": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "polkit-pkla-compat",
                    "release": "12.el8",
                    "source": "rpm",
                    "version": "0.1"
                }
            ],
            "popt": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "popt",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "1.18"
                }
            ],
            "prefixdevname": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "prefixdevname",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "0.1.0"
                }
            ],
            "procps-ng": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "procps-ng",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "3.3.15"
                }
            ],
            "psmisc": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "psmisc",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "23.1"
                }
            ],
            "publicsuffix-list-dafsa": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "publicsuffix-list-dafsa",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "20180723"
                }
            ],
            "python3-audit": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-audit",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "3.1.2"
                }
            ],
            "python3-babel": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-babel",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "2.5.1"
                }
            ],
            "python3-blivet": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "python3-blivet",
                    "release": "8.el8",
                    "source": "rpm",
                    "version": "3.6.0"
                }
            ],
            "python3-blockdev": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-blockdev",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "python3-bytesize": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-bytesize",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "1.4"
                }
            ],
            "python3-cffi": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-cffi",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "1.11.5"
                }
            ],
            "python3-chardet": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-chardet",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "3.0.4"
                }
            ],
            "python3-configobj": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-configobj",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "5.0.6"
                }
            ],
            "python3-configshell": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "python3-configshell",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "1.1.28"
                }
            ],
            "python3-cryptography": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-cryptography",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "3.2.1"
                }
            ],
            "python3-dateutil": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "python3-dateutil",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "2.6.1"
                }
            ],
            "python3-dbus": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-dbus",
                    "release": "15.el8",
                    "source": "rpm",
                    "version": "1.2.4"
                }
            ],
            "python3-dbus-client-gen": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-dbus-client-gen",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "0.4"
                }
            ],
            "python3-dbus-python-client-gen": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-dbus-python-client-gen",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "0.7"
                }
            ],
            "python3-dbus-signature-pyparsing": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-dbus-signature-pyparsing",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.03"
                }
            ],
            "python3-decorator": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-decorator",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "4.2.1"
                }
            ],
            "python3-dnf": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-dnf",
                    "release": "20.el8",
                    "source": "rpm",
                    "version": "4.7.0"
                }
            ],
            "python3-dnf-plugins-core": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-dnf-plugins-core",
                    "release": "25.el8",
                    "source": "rpm",
                    "version": "4.0.21"
                }
            ],
            "python3-firewall": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-firewall",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "0.9.11"
                }
            ],
            "python3-gobject-base": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-gobject-base",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "3.28.3"
                }
            ],
            "python3-gpg": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-gpg",
                    "release": "12.el8",
                    "source": "rpm",
                    "version": "1.13.1"
                }
            ],
            "python3-hawkey": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-hawkey",
                    "release": "19.el8",
                    "source": "rpm",
                    "version": "0.63.0"
                }
            ],
            "python3-html5lib": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "python3-html5lib",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "0.999999999"
                }
            ],
            "python3-idna": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-idna",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "2.5"
                }
            ],
            "python3-into-dbus-python": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-into-dbus-python",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.06"
                }
            ],
            "python3-jinja2": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-jinja2",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "2.10.1"
                }
            ],
            "python3-jsonpatch": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-jsonpatch",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "1.21"
                }
            ],
            "python3-jsonpointer": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-jsonpointer",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "1.10"
                }
            ],
            "python3-jsonschema": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-jsonschema",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "2.6.0"
                }
            ],
            "python3-justbases": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-justbases",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "0.14"
                }
            ],
            "python3-justbytes": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-justbytes",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.14"
                }
            ],
            "python3-jwt": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-jwt",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "1.6.1"
                }
            ],
            "python3-kmod": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-kmod",
                    "release": "20.el8",
                    "source": "rpm",
                    "version": "0.9"
                }
            ],
            "python3-libcomps": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-libcomps",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "0.1.18"
                }
            ],
            "python3-libdnf": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-libdnf",
                    "release": "19.el8",
                    "source": "rpm",
                    "version": "0.63.0"
                }
            ],
            "python3-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-libs",
                    "release": "62.el8",
                    "source": "rpm",
                    "version": "3.6.8"
                }
            ],
            "python3-libselinux": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-libselinux",
                    "release": "8.el8",
                    "source": "rpm",
                    "version": "2.9"
                }
            ],
            "python3-libsemanage": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-libsemanage",
                    "release": "9.el8",
                    "source": "rpm",
                    "version": "2.9"
                }
            ],
            "python3-linux-procfs": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-linux-procfs",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "0.7.3"
                }
            ],
            "python3-lxml": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-lxml",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "4.2.3"
                }
            ],
            "python3-markupsafe": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-markupsafe",
                    "release": "19.el8",
                    "source": "rpm",
                    "version": "0.23"
                }
            ],
            "python3-netifaces": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-netifaces",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "0.10.6"
                }
            ],
            "python3-nftables": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "python3-nftables",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "1.0.4"
                }
            ],
            "python3-oauthlib": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-oauthlib",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "2.1.0"
                }
            ],
            "python3-perf": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-perf",
                    "release": "553.5.1.el8",
                    "source": "rpm",
                    "version": "4.18.0"
                }
            ],
            "python3-pip": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-pip",
                    "release": "24.el8",
                    "source": "rpm",
                    "version": "9.0.3"
                }
            ],
            "python3-pip-wheel": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-pip-wheel",
                    "release": "24.el8",
                    "source": "rpm",
                    "version": "9.0.3"
                }
            ],
            "python3-ply": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-ply",
                    "release": "9.el8",
                    "source": "rpm",
                    "version": "3.9"
                }
            ],
            "python3-policycoreutils": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-policycoreutils",
                    "release": "26.el8",
                    "source": "rpm",
                    "version": "2.9"
                }
            ],
            "python3-prettytable": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-prettytable",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "0.7.2"
                }
            ],
            "python3-psutil": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-psutil",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "5.4.3"
                }
            ],
            "python3-pycparser": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-pycparser",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "2.14"
                }
            ],
            "python3-pyparsing": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-pyparsing",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "2.1.10"
                }
            ],
            "python3-pyparted": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "python3-pyparted",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "3.11.7"
                }
            ],
            "python3-pyserial": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-pyserial",
                    "release": "9.el8",
                    "source": "rpm",
                    "version": "3.1.1"
                }
            ],
            "python3-pysocks": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-pysocks",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "1.6.8"
                }
            ],
            "python3-pytz": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-pytz",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "2017.2"
                }
            ],
            "python3-pyudev": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-pyudev",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "0.21.0"
                }
            ],
            "python3-pyyaml": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-pyyaml",
                    "release": "12.el8",
                    "source": "rpm",
                    "version": "3.12"
                }
            ],
            "python3-requests": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-requests",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "2.20.0"
                }
            ],
            "python3-rpm": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-rpm",
                    "release": "31.el8",
                    "source": "rpm",
                    "version": "4.14.3"
                }
            ],
            "python3-rtslib": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-rtslib",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "2.1.75"
                }
            ],
            "python3-setools": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-setools",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "4.3.0"
                }
            ],
            "python3-setuptools": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-setuptools",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "39.2.0"
                }
            ],
            "python3-setuptools-wheel": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-setuptools-wheel",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "39.2.0"
                }
            ],
            "python3-six": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-six",
                    "release": "8.el8",
                    "source": "rpm",
                    "version": "1.11.0"
                }
            ],
            "python3-slip": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-slip",
                    "release": "13.el8",
                    "source": "rpm",
                    "version": "0.6.4"
                }
            ],
            "python3-slip-dbus": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-slip-dbus",
                    "release": "13.el8",
                    "source": "rpm",
                    "version": "0.6.4"
                }
            ],
            "python3-syspurpose": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-syspurpose",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "1.28.42"
                }
            ],
            "python3-systemd": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-systemd",
                    "release": "8.el8",
                    "source": "rpm",
                    "version": "234"
                }
            ],
            "python3-unbound": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-unbound",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "1.16.2"
                }
            ],
            "python3-urllib3": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-urllib3",
                    "release": "7.el8",
                    "source": "rpm",
                    "version": "1.24.2"
                }
            ],
            "python3-urwid": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python3-urwid",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "1.3.1"
                }
            ],
            "python3-webencodings": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "python3-webencodings",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "0.5.1"
                }
            ],
            "python36": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "python36",
                    "release": "39.module_el8+762+77bd8591",
                    "source": "rpm",
                    "version": "3.6.8"
                }
            ],
            "qa-tools": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "qa-tools",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "4.1"
                }
            ],
            "qemu-guest-agent": [
                {
                    "arch": "x86_64",
                    "epoch": 15,
                    "name": "qemu-guest-agent",
                    "release": "49.module_el8+991+097e156d",
                    "source": "rpm",
                    "version": "6.2.0"
                }
            ],
            "quota": [
                {
                    "arch": "x86_64",
                    "epoch": 1,
                    "name": "quota",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "4.04"
                }
            ],
            "quota-nls": [
                {
                    "arch": "noarch",
                    "epoch": 1,
                    "name": "quota-nls",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "4.04"
                }
            ],
            "readline": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "readline",
                    "release": "10.el8",
                    "source": "rpm",
                    "version": "7.0"
                }
            ],
            "restraint": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "restraint",
                    "release": "1.el8bkr",
                    "source": "rpm",
                    "version": "0.4.4"
                }
            ],
            "restraint-rhts": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "restraint-rhts",
                    "release": "1.el8bkr",
                    "source": "rpm",
                    "version": "0.4.4"
                }
            ],
            "rng-tools": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "rng-tools",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "6.16"
                }
            ],
            "rootfiles": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "rootfiles",
                    "release": "22.el8",
                    "source": "rpm",
                    "version": "8.1"
                }
            ],
            "rpcbind": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "rpcbind",
                    "release": "10.el8",
                    "source": "rpm",
                    "version": "1.2.5"
                }
            ],
            "rpm": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "rpm",
                    "release": "31.el8",
                    "source": "rpm",
                    "version": "4.14.3"
                }
            ],
            "rpm-build-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "rpm-build-libs",
                    "release": "31.el8",
                    "source": "rpm",
                    "version": "4.14.3"
                }
            ],
            "rpm-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "rpm-libs",
                    "release": "31.el8",
                    "source": "rpm",
                    "version": "4.14.3"
                }
            ],
            "rpm-plugin-selinux": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "rpm-plugin-selinux",
                    "release": "31.el8",
                    "source": "rpm",
                    "version": "4.14.3"
                }
            ],
            "rpm-plugin-systemd-inhibit": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "rpm-plugin-systemd-inhibit",
                    "release": "31.el8",
                    "source": "rpm",
                    "version": "4.14.3"
                }
            ],
            "rsync": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "rsync",
                    "release": "19.el8.1",
                    "source": "rpm",
                    "version": "3.1.3"
                }
            ],
            "rsyslog": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "rsyslog",
                    "release": "15.el8",
                    "source": "rpm",
                    "version": "8.2102.0"
                }
            ],
            "sed": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "sed",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "4.5"
                }
            ],
            "selinux-policy": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "selinux-policy",
                    "release": "139.el8",
                    "source": "rpm",
                    "version": "3.14.3"
                }
            ],
            "selinux-policy-targeted": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "selinux-policy-targeted",
                    "release": "139.el8",
                    "source": "rpm",
                    "version": "3.14.3"
                }
            ],
            "setup": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "setup",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "2.12.2"
                }
            ],
            "sg3_utils": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "sg3_utils",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "1.44"
                }
            ],
            "sg3_utils-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "sg3_utils-libs",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "1.44"
                }
            ],
            "shadow-utils": [
                {
                    "arch": "x86_64",
                    "epoch": 2,
                    "name": "shadow-utils",
                    "release": "22.el8",
                    "source": "rpm",
                    "version": "4.6"
                }
            ],
            "shared-mime-info": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "shared-mime-info",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "1.9"
                }
            ],
            "slang": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "slang",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.3.2"
                }
            ],
            "snappy": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "snappy",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "1.1.8"
                }
            ],
            "sqlite-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "sqlite-libs",
                    "release": "19.el8",
                    "source": "rpm",
                    "version": "3.26.0"
                }
            ],
            "squashfs-tools": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "squashfs-tools",
                    "release": "21.el8",
                    "source": "rpm",
                    "version": "4.3"
                }
            ],
            "sssd-client": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "sssd-client",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.9.4"
                }
            ],
            "sssd-common": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "sssd-common",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.9.4"
                }
            ],
            "sssd-kcm": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "sssd-kcm",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.9.4"
                }
            ],
            "sssd-nfs-idmap": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "sssd-nfs-idmap",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "2.9.4"
                }
            ],
            "strace": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "strace",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "5.18"
                }
            ],
            "stratis-cli": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "stratis-cli",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "2.4.2"
                }
            ],
            "stratisd": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "stratisd",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "2.4.2"
                }
            ],
            "sudo": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "sudo",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "1.9.5p2"
                }
            ],
            "systemd": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "systemd",
                    "release": "82.el8.1",
                    "source": "rpm",
                    "version": "239"
                }
            ],
            "systemd-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "systemd-libs",
                    "release": "82.el8.1",
                    "source": "rpm",
                    "version": "239"
                }
            ],
            "systemd-pam": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "systemd-pam",
                    "release": "82.el8.1",
                    "source": "rpm",
                    "version": "239"
                }
            ],
            "systemd-udev": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "systemd-udev",
                    "release": "82.el8.1",
                    "source": "rpm",
                    "version": "239"
                }
            ],
            "systemtap": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "systemtap",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "4.9"
                }
            ],
            "systemtap-client": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "systemtap-client",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "4.9"
                }
            ],
            "systemtap-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "systemtap-devel",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "4.9"
                }
            ],
            "systemtap-runtime": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "systemtap-runtime",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "4.9"
                }
            ],
            "tar": [
                {
                    "arch": "x86_64",
                    "epoch": 2,
                    "name": "tar",
                    "release": "9.el8",
                    "source": "rpm",
                    "version": "1.30"
                }
            ],
            "target-restore": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "target-restore",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "2.1.75"
                }
            ],
            "targetcli": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "targetcli",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "2.1.53"
                }
            ],
            "tbb": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "tbb",
                    "release": "9.el8",
                    "source": "rpm",
                    "version": "2018.2"
                }
            ],
            "teamd": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "teamd",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "1.31"
                }
            ],
            "time": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "time",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "1.9"
                }
            ],
            "timedatex": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "timedatex",
                    "release": "3.el8",
                    "source": "rpm",
                    "version": "0.5"
                }
            ],
            "tpm2-tools": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "tpm2-tools",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "4.1.1"
                }
            ],
            "tpm2-tss": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "tpm2-tss",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "2.3.2"
                }
            ],
            "trousers": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "trousers",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.3.15"
                }
            ],
            "trousers-lib": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "trousers-lib",
                    "release": "2.el8",
                    "source": "rpm",
                    "version": "0.3.15"
                }
            ],
            "tuned": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "tuned",
                    "release": "4.el8.1",
                    "source": "rpm",
                    "version": "2.22.1"
                }
            ],
            "tzdata": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "tzdata",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "2024a"
                }
            ],
            "unbound-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "unbound-libs",
                    "release": "5.el8",
                    "source": "rpm",
                    "version": "1.16.2"
                }
            ],
            "unzip": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "unzip",
                    "release": "46.el8",
                    "source": "rpm",
                    "version": "6.0"
                }
            ],
            "userspace-rcu": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "userspace-rcu",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "0.10.1"
                }
            ],
            "util-linux": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "util-linux",
                    "release": "46.el8",
                    "source": "rpm",
                    "version": "2.32.1"
                }
            ],
            "vdo": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "vdo",
                    "release": "14.el8",
                    "source": "rpm",
                    "version": "6.2.9.7"
                }
            ],
            "vim-common": [
                {
                    "arch": "x86_64",
                    "epoch": 2,
                    "name": "vim-common",
                    "release": "19.el8.4",
                    "source": "rpm",
                    "version": "8.0.1763"
                }
            ],
            "vim-enhanced": [
                {
                    "arch": "x86_64",
                    "epoch": 2,
                    "name": "vim-enhanced",
                    "release": "19.el8.4",
                    "source": "rpm",
                    "version": "8.0.1763"
                }
            ],
            "vim-filesystem": [
                {
                    "arch": "noarch",
                    "epoch": 2,
                    "name": "vim-filesystem",
                    "release": "19.el8.4",
                    "source": "rpm",
                    "version": "8.0.1763"
                }
            ],
            "vim-minimal": [
                {
                    "arch": "x86_64",
                    "epoch": 2,
                    "name": "vim-minimal",
                    "release": "19.el8.4",
                    "source": "rpm",
                    "version": "8.0.1763"
                }
            ],
            "virt-what": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "virt-what",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "1.25"
                }
            ],
            "volume_key-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "volume_key-libs",
                    "release": "6.el8",
                    "source": "rpm",
                    "version": "0.3.11"
                }
            ],
            "wget": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "wget",
                    "release": "11.el8",
                    "source": "rpm",
                    "version": "1.19.5"
                }
            ],
            "which": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "which",
                    "release": "20.el8",
                    "source": "rpm",
                    "version": "2.21"
                }
            ],
            "xfsprogs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "xfsprogs",
                    "release": "12.el8",
                    "source": "rpm",
                    "version": "5.0.0"
                }
            ],
            "xkeyboard-config": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "xkeyboard-config",
                    "release": "1.el8",
                    "source": "rpm",
                    "version": "2.28"
                }
            ],
            "xz": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "xz",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "5.2.4"
                }
            ],
            "xz-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "xz-devel",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "5.2.4"
                }
            ],
            "xz-libs": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "xz-libs",
                    "release": "4.el8",
                    "source": "rpm",
                    "version": "5.2.4"
                }
            ],
            "yum": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "yum",
                    "release": "20.el8",
                    "source": "rpm",
                    "version": "4.7.0"
                }
            ],
            "yum-utils": [
                {
                    "arch": "noarch",
                    "epoch": null,
                    "name": "yum-utils",
                    "release": "25.el8",
                    "source": "rpm",
                    "version": "4.0.21"
                }
            ],
            "zip": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "zip",
                    "release": "23.el8",
                    "source": "rpm",
                    "version": "3.0"
                }
            ],
            "zlib": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "zlib",
                    "release": "25.el8",
                    "source": "rpm",
                    "version": "1.2.11"
                }
            ],
            "zlib-devel": [
                {
                    "arch": "x86_64",
                    "epoch": null,
                    "name": "zlib-devel",
                    "release": "25.el8",
                    "source": "rpm",
                    "version": "1.2.11"
                }
            ]
        }
    },
    "changed": false
}

TASK [Set blivet package name] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:321
Wednesday 11 December 2024  10:12:53 -0500 (0:00:01.917)       0:05:27.759 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "blivet_pkg_name": [
            "python3-blivet"
        ]
    },
    "changed": false
}

TASK [Set blivet package version] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:325
Wednesday 11 December 2024  10:12:53 -0500 (0:00:00.055)       0:05:27.814 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "blivet_pkg_version": "3.6.0-8.el8"
    },
    "changed": false
}

TASK [Set distribution version] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:329
Wednesday 11 December 2024  10:12:53 -0500 (0:00:00.052)       0:05:27.866 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "is_fedora": false,
        "is_rhel10": false,
        "is_rhel8": true,
        "is_rhel9": false
    },
    "changed": false
}

TASK [Create one LVM logical volume under one volume group with size 5g] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:341
Wednesday 11 December 2024  10:12:53 -0500 (0:00:00.080)       0:05:27.947 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:12:53 -0500 (0:00:00.068)       0:05:28.015 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:12:53 -0500 (0:00:00.073)       0:05:28.088 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:12:53 -0500 (0:00:00.053)       0:05:28.141 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:12:53 -0500 (0:00:00.112)       0:05:28.253 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:12:54 -0500 (0:00:00.043)       0:05:28.297 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:12:54 -0500 (0:00:00.044)       0:05:28.342 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:12:54 -0500 (0:00:00.044)       0:05:28.387 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:12:54 -0500 (0:00:00.045)       0:05:28.432 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:12:54 -0500 (0:00:00.144)       0:05:28.577 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:12:54 -0500 (0:00:00.052)       0:05:28.629 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "5g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:12:54 -0500 (0:00:00.052)       0:05:28.681 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:12:54 -0500 (0:00:00.047)       0:05:28.729 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:12:54 -0500 (0:00:00.045)       0:05:28.774 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:12:54 -0500 (0:00:00.046)       0:05:28.820 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:12:54 -0500 (0:00:00.043)       0:05:28.864 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:12:54 -0500 (0:00:00.045)       0:05:28.909 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:12:54 -0500 (0:00:00.062)       0:05:28.972 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:12:54 -0500 (0:00:00.038)       0:05:29.010 **** 
changed: [managed-node3] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/sda",
            "fs_type": "lvmpv"
        },
        {
            "action": "create device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "create device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1",
        "/dev/mapper/foo-test1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "lvm2",
        "xfsprogs",
        "e2fsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:12:59 -0500 (0:00:04.701)       0:05:33.711 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:12:59 -0500 (0:00:00.059)       0:05:33.771 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929959.0386522,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "ab8070345774adad92683e9645714452be7be474",
        "ctime": 1733929958.500654,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929958.500654,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1343,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:12:59 -0500 (0:00:00.415)       0:05:34.187 **** 
ok: [managed-node3] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:13:00 -0500 (0:00:00.424)       0:05:34.611 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:13:00 -0500 (0:00:00.037)       0:05:34.648 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/sda",
                "fs_type": "lvmpv"
            },
            {
                "action": "create device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "create device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1",
            "/dev/mapper/foo-test1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "lvm2",
            "xfsprogs",
            "e2fsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:13:00 -0500 (0:00:00.098)       0:05:34.747 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:13:00 -0500 (0:00:00.054)       0:05:34.801 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:13:00 -0500 (0:00:00.045)       0:05:34.847 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:13:00 -0500 (0:00:00.039)       0:05:34.887 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:13:01 -0500 (0:00:00.675)       0:05:35.562 **** 
changed: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:13:01 -0500 (0:00:00.465)       0:05:36.028 **** 
skipping: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:13:01 -0500 (0:00:00.066)       0:05:36.094 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:13:02 -0500 (0:00:00.732)       0:05:36.826 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:13:02 -0500 (0:00:00.454)       0:05:37.280 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:13:03 -0500 (0:00:00.040)       0:05:37.321 **** 
ok: [managed-node3]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:357
Wednesday 11 December 2024  10:13:03 -0500 (0:00:00.838)       0:05:38.159 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node3

TASK [Print out pool information] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Wednesday 11 December 2024  10:13:03 -0500 (0:00:00.101)       0:05:38.260 **** 
ok: [managed-node3] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Wednesday 11 December 2024  10:13:04 -0500 (0:00:00.058)       0:05:38.319 **** 
skipping: [managed-node3] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Wednesday 11 December 2024  10:13:04 -0500 (0:00:00.052)       0:05:38.371 **** 
ok: [managed-node3] => {
    "changed": false,
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "ext4",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/foo-test1",
            "size": "5G",
            "type": "lvm",
            "uuid": "92b3e76d-b88c-4b0e-beef-1b86c5c4e02a"
        },
        "/dev/sda": {
            "fstype": "LVM2_member",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "3yxKep-6iZX-CsAB-6nEn-xjuh-32fU-otZ227"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Wednesday 11 December 2024  10:13:04 -0500 (0:00:00.404)       0:05:38.775 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003316",
    "end": "2024-12-11 10:13:04.798204",
    "rc": 0,
    "start": "2024-12-11 10:13:04.794888"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Wednesday 11 December 2024  10:13:04 -0500 (0:00:00.415)       0:05:39.191 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002690",
    "end": "2024-12-11 10:13:05.253544",
    "failed_when_result": false,
    "rc": 0,
    "start": "2024-12-11 10:13:05.250854"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Wednesday 11 December 2024  10:13:05 -0500 (0:00:00.432)       0:05:39.624 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node3

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Wednesday 11 December 2024  10:13:05 -0500 (0:00:00.080)       0:05:39.705 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Wednesday 11 December 2024  10:13:05 -0500 (0:00:00.044)       0:05:39.750 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "vgs",
        "--noheadings",
        "--binary",
        "-o",
        "shared",
        "foo"
    ],
    "delta": "0:00:00.023810",
    "end": "2024-12-11 10:13:05.844014",
    "rc": 0,
    "start": "2024-12-11 10:13:05.820204"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Wednesday 11 December 2024  10:13:05 -0500 (0:00:00.473)       0:05:40.223 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Wednesday 11 December 2024  10:13:05 -0500 (0:00:00.052)       0:05:40.276 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Wednesday 11 December 2024  10:13:06 -0500 (0:00:00.090)       0:05:40.367 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Wednesday 11 December 2024  10:13:06 -0500 (0:00:00.052)       0:05:40.419 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sda",
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Wednesday 11 December 2024  10:13:06 -0500 (0:00:00.401)       0:05:40.820 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Wednesday 11 December 2024  10:13:06 -0500 (0:00:00.047)       0:05:40.867 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Wednesday 11 December 2024  10:13:06 -0500 (0:00:00.108)       0:05:40.976 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Wednesday 11 December 2024  10:13:06 -0500 (0:00:00.052)       0:05:41.028 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Wednesday 11 December 2024  10:13:06 -0500 (0:00:00.047)       0:05:41.075 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Wednesday 11 December 2024  10:13:06 -0500 (0:00:00.049)       0:05:41.125 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Wednesday 11 December 2024  10:13:06 -0500 (0:00:00.044)       0:05:41.169 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Wednesday 11 December 2024  10:13:06 -0500 (0:00:00.056)       0:05:41.226 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.41.57 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Wednesday 11 December 2024  10:13:07 -0500 (0:00:00.444)       0:05:41.671 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Wednesday 11 December 2024  10:13:07 -0500 (0:00:00.046)       0:05:41.717 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node3

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Wednesday 11 December 2024  10:13:07 -0500 (0:00:00.085)       0:05:41.802 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Wednesday 11 December 2024  10:13:07 -0500 (0:00:00.044)       0:05:41.847 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Wednesday 11 December 2024  10:13:07 -0500 (0:00:00.042)       0:05:41.890 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Wednesday 11 December 2024  10:13:07 -0500 (0:00:00.053)       0:05:41.944 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Wednesday 11 December 2024  10:13:07 -0500 (0:00:00.101)       0:05:42.046 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Wednesday 11 December 2024  10:13:07 -0500 (0:00:00.046)       0:05:42.092 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Wednesday 11 December 2024  10:13:07 -0500 (0:00:00.044)       0:05:42.136 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Wednesday 11 December 2024  10:13:07 -0500 (0:00:00.044)       0:05:42.181 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Wednesday 11 December 2024  10:13:07 -0500 (0:00:00.043)       0:05:42.225 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Wednesday 11 December 2024  10:13:07 -0500 (0:00:00.042)       0:05:42.267 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Wednesday 11 December 2024  10:13:08 -0500 (0:00:00.043)       0:05:42.311 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Wednesday 11 December 2024  10:13:08 -0500 (0:00:00.046)       0:05:42.358 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node3

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Wednesday 11 December 2024  10:13:08 -0500 (0:00:00.084)       0:05:42.443 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node3

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Wednesday 11 December 2024  10:13:08 -0500 (0:00:00.079)       0:05:42.522 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Wednesday 11 December 2024  10:13:08 -0500 (0:00:00.043)       0:05:42.566 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Wednesday 11 December 2024  10:13:08 -0500 (0:00:00.043)       0:05:42.609 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Wednesday 11 December 2024  10:13:08 -0500 (0:00:00.099)       0:05:42.708 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Wednesday 11 December 2024  10:13:08 -0500 (0:00:00.044)       0:05:42.753 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Wednesday 11 December 2024  10:13:08 -0500 (0:00:00.045)       0:05:42.798 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Wednesday 11 December 2024  10:13:08 -0500 (0:00:00.051)       0:05:42.849 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Wednesday 11 December 2024  10:13:08 -0500 (0:00:00.063)       0:05:42.913 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node3

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Wednesday 11 December 2024  10:13:08 -0500 (0:00:00.124)       0:05:43.038 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node3

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Wednesday 11 December 2024  10:13:08 -0500 (0:00:00.120)       0:05:43.159 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Wednesday 11 December 2024  10:13:08 -0500 (0:00:00.051)       0:05:43.210 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Wednesday 11 December 2024  10:13:08 -0500 (0:00:00.049)       0:05:43.259 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Wednesday 11 December 2024  10:13:09 -0500 (0:00:00.044)       0:05:43.304 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Wednesday 11 December 2024  10:13:09 -0500 (0:00:00.045)       0:05:43.350 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Wednesday 11 December 2024  10:13:09 -0500 (0:00:00.088)       0:05:43.438 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Wednesday 11 December 2024  10:13:09 -0500 (0:00:00.103)       0:05:43.542 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Wednesday 11 December 2024  10:13:09 -0500 (0:00:00.050)       0:05:43.593 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node3

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Wednesday 11 December 2024  10:13:09 -0500 (0:00:00.081)       0:05:43.674 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Wednesday 11 December 2024  10:13:09 -0500 (0:00:00.047)       0:05:43.722 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Wednesday 11 December 2024  10:13:09 -0500 (0:00:00.050)       0:05:43.772 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Wednesday 11 December 2024  10:13:09 -0500 (0:00:00.044)       0:05:43.816 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Wednesday 11 December 2024  10:13:09 -0500 (0:00:00.044)       0:05:43.860 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Wednesday 11 December 2024  10:13:09 -0500 (0:00:00.042)       0:05:43.903 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Wednesday 11 December 2024  10:13:09 -0500 (0:00:00.045)       0:05:43.948 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Wednesday 11 December 2024  10:13:09 -0500 (0:00:00.046)       0:05:43.995 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node3

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Wednesday 11 December 2024  10:13:09 -0500 (0:00:00.102)       0:05:44.097 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node3

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Wednesday 11 December 2024  10:13:09 -0500 (0:00:00.134)       0:05:44.232 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Wednesday 11 December 2024  10:13:09 -0500 (0:00:00.044)       0:05:44.276 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Wednesday 11 December 2024  10:13:10 -0500 (0:00:00.046)       0:05:44.323 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Wednesday 11 December 2024  10:13:10 -0500 (0:00:00.043)       0:05:44.366 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Wednesday 11 December 2024  10:13:10 -0500 (0:00:00.044)       0:05:44.411 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Wednesday 11 December 2024  10:13:10 -0500 (0:00:00.045)       0:05:44.456 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Wednesday 11 December 2024  10:13:10 -0500 (0:00:00.044)       0:05:44.501 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Wednesday 11 December 2024  10:13:10 -0500 (0:00:00.045)       0:05:44.546 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node3

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Wednesday 11 December 2024  10:13:10 -0500 (0:00:00.093)       0:05:44.639 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Wednesday 11 December 2024  10:13:10 -0500 (0:00:00.050)       0:05:44.690 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Wednesday 11 December 2024  10:13:10 -0500 (0:00:00.062)       0:05:44.753 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Wednesday 11 December 2024  10:13:10 -0500 (0:00:00.067)       0:05:44.820 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Wednesday 11 December 2024  10:13:10 -0500 (0:00:00.056)       0:05:44.877 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Wednesday 11 December 2024  10:13:10 -0500 (0:00:00.118)       0:05:44.995 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Wednesday 11 December 2024  10:13:10 -0500 (0:00:00.051)       0:05:45.046 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Wednesday 11 December 2024  10:13:10 -0500 (0:00:00.054)       0:05:45.100 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node3

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Wednesday 11 December 2024  10:13:10 -0500 (0:00:00.080)       0:05:45.181 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Wednesday 11 December 2024  10:13:10 -0500 (0:00:00.049)       0:05:45.230 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node3

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Wednesday 11 December 2024  10:13:11 -0500 (0:00:00.197)       0:05:45.427 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Wednesday 11 December 2024  10:13:11 -0500 (0:00:00.066)       0:05:45.494 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Wednesday 11 December 2024  10:13:11 -0500 (0:00:00.055)       0:05:45.550 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Wednesday 11 December 2024  10:13:11 -0500 (0:00:00.052)       0:05:45.602 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Wednesday 11 December 2024  10:13:11 -0500 (0:00:00.054)       0:05:45.657 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Wednesday 11 December 2024  10:13:11 -0500 (0:00:00.046)       0:05:45.703 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Wednesday 11 December 2024  10:13:11 -0500 (0:00:00.103)       0:05:45.807 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Wednesday 11 December 2024  10:13:11 -0500 (0:00:00.049)       0:05:45.857 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Wednesday 11 December 2024  10:13:11 -0500 (0:00:00.058)       0:05:45.915 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Wednesday 11 December 2024  10:13:11 -0500 (0:00:00.067)       0:05:45.983 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Wednesday 11 December 2024  10:13:11 -0500 (0:00:00.068)       0:05:46.052 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Wednesday 11 December 2024  10:13:11 -0500 (0:00:00.087)       0:05:46.139 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Wednesday 11 December 2024  10:13:11 -0500 (0:00:00.097)       0:05:46.237 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Wednesday 11 December 2024  10:13:11 -0500 (0:00:00.059)       0:05:46.296 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Wednesday 11 December 2024  10:13:12 -0500 (0:00:00.061)       0:05:46.357 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Wednesday 11 December 2024  10:13:12 -0500 (0:00:00.053)       0:05:46.411 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Wednesday 11 December 2024  10:13:12 -0500 (0:00:00.054)       0:05:46.465 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Wednesday 11 December 2024  10:13:12 -0500 (0:00:00.046)       0:05:46.512 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Wednesday 11 December 2024  10:13:12 -0500 (0:00:00.053)       0:05:46.565 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Wednesday 11 December 2024  10:13:12 -0500 (0:00:00.058)       0:05:46.623 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929979.255586,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1733929979.255586,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 372868,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1733929979.255586,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Wednesday 11 December 2024  10:13:12 -0500 (0:00:00.488)       0:05:47.112 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Wednesday 11 December 2024  10:13:12 -0500 (0:00:00.065)       0:05:47.177 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Wednesday 11 December 2024  10:13:12 -0500 (0:00:00.045)       0:05:47.223 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Wednesday 11 December 2024  10:13:12 -0500 (0:00:00.049)       0:05:47.273 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Wednesday 11 December 2024  10:13:13 -0500 (0:00:00.045)       0:05:47.319 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Wednesday 11 December 2024  10:13:13 -0500 (0:00:00.045)       0:05:47.364 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Wednesday 11 December 2024  10:13:13 -0500 (0:00:00.055)       0:05:47.419 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Wednesday 11 December 2024  10:13:13 -0500 (0:00:00.072)       0:05:47.492 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Wednesday 11 December 2024  10:13:16 -0500 (0:00:02.882)       0:05:50.374 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Wednesday 11 December 2024  10:13:16 -0500 (0:00:00.049)       0:05:50.423 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Wednesday 11 December 2024  10:13:16 -0500 (0:00:00.064)       0:05:50.488 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Wednesday 11 December 2024  10:13:16 -0500 (0:00:00.064)       0:05:50.552 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Wednesday 11 December 2024  10:13:16 -0500 (0:00:00.066)       0:05:50.619 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Wednesday 11 December 2024  10:13:16 -0500 (0:00:00.070)       0:05:50.689 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Wednesday 11 December 2024  10:13:16 -0500 (0:00:00.055)       0:05:50.745 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Wednesday 11 December 2024  10:13:16 -0500 (0:00:00.051)       0:05:50.796 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Wednesday 11 December 2024  10:13:16 -0500 (0:00:00.045)       0:05:50.842 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Wednesday 11 December 2024  10:13:16 -0500 (0:00:00.050)       0:05:50.892 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Wednesday 11 December 2024  10:13:16 -0500 (0:00:00.049)       0:05:50.942 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Wednesday 11 December 2024  10:13:16 -0500 (0:00:00.106)       0:05:51.048 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Wednesday 11 December 2024  10:13:16 -0500 (0:00:00.066)       0:05:51.115 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Wednesday 11 December 2024  10:13:16 -0500 (0:00:00.070)       0:05:51.185 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Wednesday 11 December 2024  10:13:16 -0500 (0:00:00.073)       0:05:51.258 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Wednesday 11 December 2024  10:13:17 -0500 (0:00:00.081)       0:05:51.340 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Wednesday 11 December 2024  10:13:17 -0500 (0:00:00.067)       0:05:51.407 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Wednesday 11 December 2024  10:13:17 -0500 (0:00:00.073)       0:05:51.481 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Wednesday 11 December 2024  10:13:17 -0500 (0:00:00.071)       0:05:51.553 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Wednesday 11 December 2024  10:13:17 -0500 (0:00:00.069)       0:05:51.623 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Wednesday 11 December 2024  10:13:17 -0500 (0:00:00.069)       0:05:51.692 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Wednesday 11 December 2024  10:13:17 -0500 (0:00:00.068)       0:05:51.761 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Wednesday 11 December 2024  10:13:17 -0500 (0:00:00.064)       0:05:51.825 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Wednesday 11 December 2024  10:13:17 -0500 (0:00:00.069)       0:05:51.894 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Wednesday 11 December 2024  10:13:17 -0500 (0:00:00.067)       0:05:51.962 **** 
ok: [managed-node3] => {
    "bytes": 5368709120,
    "changed": false,
    "lvm": "5g",
    "parted": "5GiB",
    "size": "5 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Wednesday 11 December 2024  10:13:18 -0500 (0:00:00.483)       0:05:52.445 **** 
ok: [managed-node3] => {
    "bytes": 5368709120,
    "changed": false,
    "lvm": "5g",
    "parted": "5GiB",
    "size": "5 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Wednesday 11 December 2024  10:13:18 -0500 (0:00:00.402)       0:05:52.848 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_expected_size": "5368709120"
    },
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Wednesday 11 December 2024  10:13:18 -0500 (0:00:00.053)       0:05:52.901 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Wednesday 11 December 2024  10:13:18 -0500 (0:00:00.047)       0:05:52.948 **** 
ok: [managed-node3] => {
    "bytes": 10726680821,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.406)       0:05:53.355 **** 
skipping: [managed-node3] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.047)       0:05:53.403 **** 
skipping: [managed-node3] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.048)       0:05:53.451 **** 
skipping: [managed-node3] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.048)       0:05:53.500 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.048)       0:05:53.549 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.043)       0:05:53.593 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.045)       0:05:53.638 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.044)       0:05:53.683 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.105)       0:05:53.788 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.043)       0:05:53.832 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.045)       0:05:53.877 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.045)       0:05:53.923 **** 
skipping: [managed-node3] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.046)       0:05:53.969 **** 
skipping: [managed-node3] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.043)       0:05:54.012 **** 
skipping: [managed-node3] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.045)       0:05:54.058 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.045)       0:05:54.103 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.046)       0:05:54.150 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.043)       0:05:54.193 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.045)       0:05:54.238 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Wednesday 11 December 2024  10:13:19 -0500 (0:00:00.044)       0:05:54.283 **** 
ok: [managed-node3] => {
    "storage_test_actual_size": {
        "bytes": 5368709120,
        "changed": false,
        "failed": false,
        "lvm": "5g",
        "parted": "5GiB",
        "size": "5 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Wednesday 11 December 2024  10:13:20 -0500 (0:00:00.051)       0:05:54.334 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "5368709120"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Wednesday 11 December 2024  10:13:20 -0500 (0:00:00.053)       0:05:54.388 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Wednesday 11 December 2024  10:13:20 -0500 (0:00:00.054)       0:05:54.442 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.027612",
    "end": "2024-12-11 10:13:20.557128",
    "rc": 0,
    "start": "2024-12-11 10:13:20.529516"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Wednesday 11 December 2024  10:13:20 -0500 (0:00:00.504)       0:05:54.946 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Wednesday 11 December 2024  10:13:20 -0500 (0:00:00.059)       0:05:55.006 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Wednesday 11 December 2024  10:13:20 -0500 (0:00:00.067)       0:05:55.074 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Wednesday 11 December 2024  10:13:20 -0500 (0:00:00.073)       0:05:55.148 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Wednesday 11 December 2024  10:13:20 -0500 (0:00:00.059)       0:05:55.208 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Wednesday 11 December 2024  10:13:20 -0500 (0:00:00.054)       0:05:55.262 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Wednesday 11 December 2024  10:13:21 -0500 (0:00:00.048)       0:05:55.311 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Wednesday 11 December 2024  10:13:21 -0500 (0:00:00.046)       0:05:55.357 **** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Wednesday 11 December 2024  10:13:21 -0500 (0:00:00.050)       0:05:55.407 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Change volume_size to 9g] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:360
Wednesday 11 December 2024  10:13:21 -0500 (0:00:00.049)       0:05:55.457 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:13:21 -0500 (0:00:00.107)       0:05:55.564 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:13:21 -0500 (0:00:00.113)       0:05:55.678 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:13:21 -0500 (0:00:00.249)       0:05:55.927 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:13:21 -0500 (0:00:00.180)       0:05:56.108 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:13:21 -0500 (0:00:00.058)       0:05:56.166 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:13:21 -0500 (0:00:00.050)       0:05:56.217 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:13:21 -0500 (0:00:00.046)       0:05:56.263 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:13:22 -0500 (0:00:00.047)       0:05:56.310 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:13:22 -0500 (0:00:00.100)       0:05:56.411 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:13:22 -0500 (0:00:00.049)       0:05:56.460 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "9g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:13:22 -0500 (0:00:00.054)       0:05:56.514 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:13:22 -0500 (0:00:00.070)       0:05:56.585 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:13:22 -0500 (0:00:00.051)       0:05:56.637 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:13:22 -0500 (0:00:00.065)       0:05:56.703 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:13:22 -0500 (0:00:00.083)       0:05:56.786 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:13:22 -0500 (0:00:00.131)       0:05:56.917 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:13:22 -0500 (0:00:00.091)       0:05:57.009 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:13:22 -0500 (0:00:00.062)       0:05:57.071 **** 
changed: [managed-node3] => {
    "actions": [
        {
            "action": "resize device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "resize format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/mapper/foo-test1",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "group": null,
            "mode": null,
            "opts": "defaults",
            "owner": null,
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs",
        "lvm2"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "9g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:13:27 -0500 (0:00:04.503)       0:06:01.575 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:13:27 -0500 (0:00:00.072)       0:06:01.647 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929981.6155782,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "306e713ef45e6db45ca0244c07767212954e6abb",
        "ctime": 1733929981.6125782,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929981.6125782,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1394,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:13:27 -0500 (0:00:00.590)       0:06:02.237 **** 
ok: [managed-node3] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:13:28 -0500 (0:00:00.454)       0:06:02.692 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:13:28 -0500 (0:00:00.069)       0:06:02.761 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [
            {
                "action": "resize device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "resize format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/mapper/foo-test1",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "group": null,
                "mode": null,
                "opts": "defaults",
                "owner": null,
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs",
            "lvm2"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "9g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:13:28 -0500 (0:00:00.105)       0:06:02.867 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "9g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:13:28 -0500 (0:00:00.075)       0:06:02.942 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:13:28 -0500 (0:00:00.055)       0:06:02.997 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:13:28 -0500 (0:00:00.058)       0:06:03.056 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:13:29 -0500 (0:00:00.697)       0:06:03.753 **** 
ok: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:13:29 -0500 (0:00:00.456)       0:06:04.209 **** 
skipping: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None})  => {
    "ansible_loop_var": "mount_info",
    "changed": false,
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "group": null,
        "mode": null,
        "opts": "defaults",
        "owner": null,
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:13:29 -0500 (0:00:00.075)       0:06:04.285 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:13:30 -0500 (0:00:00.738)       0:06:05.024 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:13:31 -0500 (0:00:00.471)       0:06:05.496 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:13:31 -0500 (0:00:00.059)       0:06:05.555 **** 
ok: [managed-node3]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:374
Wednesday 11 December 2024  10:13:32 -0500 (0:00:00.868)       0:06:06.424 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node3

TASK [Print out pool information] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Wednesday 11 December 2024  10:13:32 -0500 (0:00:00.122)       0:06:06.547 **** 
ok: [managed-node3] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "9g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Wednesday 11 December 2024  10:13:32 -0500 (0:00:00.078)       0:06:06.625 **** 
skipping: [managed-node3] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Wednesday 11 December 2024  10:13:32 -0500 (0:00:00.070)       0:06:06.696 **** 
ok: [managed-node3] => {
    "changed": false,
    "info": {
        "/dev/mapper/foo-test1": {
            "fstype": "ext4",
            "label": "",
            "mountpoint": "/opt/test1",
            "name": "/dev/mapper/foo-test1",
            "size": "9G",
            "type": "lvm",
            "uuid": "92b3e76d-b88c-4b0e-beef-1b86c5c4e02a"
        },
        "/dev/sda": {
            "fstype": "LVM2_member",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": "3yxKep-6iZX-CsAB-6nEn-xjuh-32fU-otZ227"
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Wednesday 11 December 2024  10:13:33 -0500 (0:00:00.657)       0:06:07.353 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002722",
    "end": "2024-12-11 10:13:33.512085",
    "rc": 0,
    "start": "2024-12-11 10:13:33.509363"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
/dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Wednesday 11 December 2024  10:13:33 -0500 (0:00:00.632)       0:06:07.986 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002646",
    "end": "2024-12-11 10:13:34.082400",
    "failed_when_result": false,
    "rc": 0,
    "start": "2024-12-11 10:13:34.079754"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Wednesday 11 December 2024  10:13:34 -0500 (0:00:00.516)       0:06:08.502 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node3

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Wednesday 11 December 2024  10:13:34 -0500 (0:00:00.163)       0:06:08.665 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Wednesday 11 December 2024  10:13:34 -0500 (0:00:00.087)       0:06:08.752 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "vgs",
        "--noheadings",
        "--binary",
        "-o",
        "shared",
        "foo"
    ],
    "delta": "0:00:00.023544",
    "end": "2024-12-11 10:13:34.871900",
    "rc": 0,
    "start": "2024-12-11 10:13:34.848356"
}

STDOUT:

        0

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Wednesday 11 December 2024  10:13:34 -0500 (0:00:00.520)       0:06:09.272 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Wednesday 11 December 2024  10:13:35 -0500 (0:00:00.070)       0:06:09.343 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Wednesday 11 December 2024  10:13:35 -0500 (0:00:00.100)       0:06:09.444 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Wednesday 11 December 2024  10:13:35 -0500 (0:00:00.051)       0:06:09.496 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/sda",
    "pv": "/dev/sda"
}

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Wednesday 11 December 2024  10:13:35 -0500 (0:00:00.450)       0:06:09.947 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Wednesday 11 December 2024  10:13:35 -0500 (0:00:00.141)       0:06:10.089 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/sda"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Wednesday 11 December 2024  10:13:35 -0500 (0:00:00.077)       0:06:10.166 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Wednesday 11 December 2024  10:13:35 -0500 (0:00:00.078)       0:06:10.245 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Wednesday 11 December 2024  10:13:36 -0500 (0:00:00.056)       0:06:10.301 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Wednesday 11 December 2024  10:13:36 -0500 (0:00:00.061)       0:06:10.363 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Wednesday 11 December 2024  10:13:36 -0500 (0:00:00.051)       0:06:10.415 **** 
ok: [managed-node3] => (item=/dev/sda) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/sda"
}

MSG:

All assertions passed

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Wednesday 11 December 2024  10:13:36 -0500 (0:00:00.055)       0:06:10.471 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.41.57 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Wednesday 11 December 2024  10:13:36 -0500 (0:00:00.452)       0:06:10.923 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "ansible_loop_var": "st_pool_pv",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "st_pool_pv": "/dev/sda"
}

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Wednesday 11 December 2024  10:13:36 -0500 (0:00:00.049)       0:06:10.973 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node3

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Wednesday 11 December 2024  10:13:36 -0500 (0:00:00.082)       0:06:11.055 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Wednesday 11 December 2024  10:13:36 -0500 (0:00:00.049)       0:06:11.105 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Wednesday 11 December 2024  10:13:36 -0500 (0:00:00.051)       0:06:11.156 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Wednesday 11 December 2024  10:13:36 -0500 (0:00:00.043)       0:06:11.200 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Wednesday 11 December 2024  10:13:37 -0500 (0:00:00.098)       0:06:11.299 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Wednesday 11 December 2024  10:13:37 -0500 (0:00:00.044)       0:06:11.343 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Wednesday 11 December 2024  10:13:37 -0500 (0:00:00.044)       0:06:11.388 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Wednesday 11 December 2024  10:13:37 -0500 (0:00:00.042)       0:06:11.431 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Wednesday 11 December 2024  10:13:37 -0500 (0:00:00.042)       0:06:11.473 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Wednesday 11 December 2024  10:13:37 -0500 (0:00:00.044)       0:06:11.518 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Wednesday 11 December 2024  10:13:37 -0500 (0:00:00.055)       0:06:11.574 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Wednesday 11 December 2024  10:13:37 -0500 (0:00:00.074)       0:06:11.648 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node3

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Wednesday 11 December 2024  10:13:37 -0500 (0:00:00.145)       0:06:11.794 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node3

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Wednesday 11 December 2024  10:13:37 -0500 (0:00:00.128)       0:06:11.923 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Wednesday 11 December 2024  10:13:37 -0500 (0:00:00.076)       0:06:12.000 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Wednesday 11 December 2024  10:13:37 -0500 (0:00:00.066)       0:06:12.067 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Wednesday 11 December 2024  10:13:37 -0500 (0:00:00.138)       0:06:12.205 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Wednesday 11 December 2024  10:13:37 -0500 (0:00:00.067)       0:06:12.272 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Wednesday 11 December 2024  10:13:38 -0500 (0:00:00.066)       0:06:12.339 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Wednesday 11 December 2024  10:13:38 -0500 (0:00:00.069)       0:06:12.408 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Wednesday 11 December 2024  10:13:38 -0500 (0:00:00.063)       0:06:12.472 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node3

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Wednesday 11 December 2024  10:13:38 -0500 (0:00:00.117)       0:06:12.590 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node3

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Wednesday 11 December 2024  10:13:38 -0500 (0:00:00.121)       0:06:12.711 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Wednesday 11 December 2024  10:13:38 -0500 (0:00:00.050)       0:06:12.762 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Wednesday 11 December 2024  10:13:38 -0500 (0:00:00.050)       0:06:12.812 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Wednesday 11 December 2024  10:13:38 -0500 (0:00:00.043)       0:06:12.855 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Wednesday 11 December 2024  10:13:38 -0500 (0:00:00.043)       0:06:12.899 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Wednesday 11 December 2024  10:13:38 -0500 (0:00:00.174)       0:06:13.073 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Wednesday 11 December 2024  10:13:38 -0500 (0:00:00.097)       0:06:13.171 **** 
skipping: [managed-node3] => (item=/dev/sda)  => {
    "_storage_test_pool_member_path": "/dev/sda",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Wednesday 11 December 2024  10:13:38 -0500 (0:00:00.089)       0:06:13.261 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed-node3

TASK [Set variables used by tests] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2
Wednesday 11 December 2024  10:13:39 -0500 (0:00:00.099)       0:06:13.361 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6
Wednesday 11 December 2024  10:13:39 -0500 (0:00:00.057)       0:06:13.419 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14
Wednesday 11 December 2024  10:13:39 -0500 (0:00:00.049)       0:06:13.468 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23
Wednesday 11 December 2024  10:13:39 -0500 (0:00:00.047)       0:06:13.516 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32
Wednesday 11 December 2024  10:13:39 -0500 (0:00:00.064)       0:06:13.581 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41
Wednesday 11 December 2024  10:13:39 -0500 (0:00:00.066)       0:06:13.647 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Wednesday 11 December 2024  10:13:39 -0500 (0:00:00.050)       0:06:13.698 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Wednesday 11 December 2024  10:13:39 -0500 (0:00:00.051)       0:06:13.749 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node3

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Wednesday 11 December 2024  10:13:39 -0500 (0:00:00.106)       0:06:13.856 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node3

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Wednesday 11 December 2024  10:13:39 -0500 (0:00:00.135)       0:06:13.992 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Wednesday 11 December 2024  10:13:39 -0500 (0:00:00.060)       0:06:14.052 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Wednesday 11 December 2024  10:13:39 -0500 (0:00:00.070)       0:06:14.123 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Wednesday 11 December 2024  10:13:39 -0500 (0:00:00.080)       0:06:14.204 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Wednesday 11 December 2024  10:13:39 -0500 (0:00:00.070)       0:06:14.275 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Wednesday 11 December 2024  10:13:40 -0500 (0:00:00.071)       0:06:14.347 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Wednesday 11 December 2024  10:13:40 -0500 (0:00:00.069)       0:06:14.416 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Wednesday 11 December 2024  10:13:40 -0500 (0:00:00.069)       0:06:14.485 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node3

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Wednesday 11 December 2024  10:13:40 -0500 (0:00:00.147)       0:06:14.632 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Wednesday 11 December 2024  10:13:40 -0500 (0:00:00.069)       0:06:14.702 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Wednesday 11 December 2024  10:13:40 -0500 (0:00:00.069)       0:06:14.771 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Wednesday 11 December 2024  10:13:40 -0500 (0:00:00.068)       0:06:14.840 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Wednesday 11 December 2024  10:13:40 -0500 (0:00:00.068)       0:06:14.908 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Wednesday 11 December 2024  10:13:40 -0500 (0:00:00.114)       0:06:15.022 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Wednesday 11 December 2024  10:13:40 -0500 (0:00:00.071)       0:06:15.094 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Wednesday 11 December 2024  10:13:40 -0500 (0:00:00.045)       0:06:15.139 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node3

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Wednesday 11 December 2024  10:13:40 -0500 (0:00:00.078)       0:06:15.218 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Wednesday 11 December 2024  10:13:40 -0500 (0:00:00.050)       0:06:15.269 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node3

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Wednesday 11 December 2024  10:13:41 -0500 (0:00:00.252)       0:06:15.521 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Wednesday 11 December 2024  10:13:41 -0500 (0:00:00.071)       0:06:15.593 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Wednesday 11 December 2024  10:13:41 -0500 (0:00:00.076)       0:06:15.669 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Wednesday 11 December 2024  10:13:41 -0500 (0:00:00.071)       0:06:15.740 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Wednesday 11 December 2024  10:13:41 -0500 (0:00:00.077)       0:06:15.818 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Wednesday 11 December 2024  10:13:41 -0500 (0:00:00.069)       0:06:15.887 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Wednesday 11 December 2024  10:13:41 -0500 (0:00:00.154)       0:06:16.042 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Wednesday 11 December 2024  10:13:41 -0500 (0:00:00.069)       0:06:16.111 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Wednesday 11 December 2024  10:13:41 -0500 (0:00:00.082)       0:06:16.194 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Wednesday 11 December 2024  10:13:41 -0500 (0:00:00.069)       0:06:16.264 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Wednesday 11 December 2024  10:13:42 -0500 (0:00:00.071)       0:06:16.335 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Wednesday 11 December 2024  10:13:42 -0500 (0:00:00.072)       0:06:16.408 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Wednesday 11 December 2024  10:13:42 -0500 (0:00:00.100)       0:06:16.508 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Wednesday 11 December 2024  10:13:42 -0500 (0:00:00.053)       0:06:16.562 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Wednesday 11 December 2024  10:13:42 -0500 (0:00:00.060)       0:06:16.622 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Wednesday 11 December 2024  10:13:42 -0500 (0:00:00.054)       0:06:16.676 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Wednesday 11 December 2024  10:13:42 -0500 (0:00:00.051)       0:06:16.728 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Wednesday 11 December 2024  10:13:42 -0500 (0:00:00.047)       0:06:16.776 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Wednesday 11 December 2024  10:13:42 -0500 (0:00:00.054)       0:06:16.830 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Wednesday 11 December 2024  10:13:42 -0500 (0:00:00.055)       0:06:16.885 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733930007.085497,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1733930007.085497,
        "dev": 6,
        "device_type": 64768,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 372868,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1733930007.085497,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Wednesday 11 December 2024  10:13:43 -0500 (0:00:00.431)       0:06:17.316 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Wednesday 11 December 2024  10:13:43 -0500 (0:00:00.078)       0:06:17.395 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Wednesday 11 December 2024  10:13:43 -0500 (0:00:00.071)       0:06:17.466 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Wednesday 11 December 2024  10:13:43 -0500 (0:00:00.076)       0:06:17.542 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Wednesday 11 December 2024  10:13:43 -0500 (0:00:00.072)       0:06:17.615 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Wednesday 11 December 2024  10:13:43 -0500 (0:00:00.051)       0:06:17.667 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Wednesday 11 December 2024  10:13:43 -0500 (0:00:00.062)       0:06:17.730 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Wednesday 11 December 2024  10:13:43 -0500 (0:00:00.051)       0:06:17.782 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Wednesday 11 December 2024  10:13:46 -0500 (0:00:02.844)       0:06:20.626 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Wednesday 11 December 2024  10:13:46 -0500 (0:00:00.043)       0:06:20.670 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Wednesday 11 December 2024  10:13:46 -0500 (0:00:00.044)       0:06:20.714 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Wednesday 11 December 2024  10:13:46 -0500 (0:00:00.053)       0:06:20.768 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Wednesday 11 December 2024  10:13:46 -0500 (0:00:00.042)       0:06:20.811 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Wednesday 11 December 2024  10:13:46 -0500 (0:00:00.047)       0:06:20.858 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Wednesday 11 December 2024  10:13:46 -0500 (0:00:00.056)       0:06:20.915 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Wednesday 11 December 2024  10:13:46 -0500 (0:00:00.066)       0:06:20.982 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Wednesday 11 December 2024  10:13:46 -0500 (0:00:00.048)       0:06:21.030 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Wednesday 11 December 2024  10:13:46 -0500 (0:00:00.062)       0:06:21.092 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Wednesday 11 December 2024  10:13:46 -0500 (0:00:00.052)       0:06:21.145 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Wednesday 11 December 2024  10:13:46 -0500 (0:00:00.121)       0:06:21.266 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Wednesday 11 December 2024  10:13:47 -0500 (0:00:00.057)       0:06:21.323 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Wednesday 11 December 2024  10:13:47 -0500 (0:00:00.050)       0:06:21.374 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Wednesday 11 December 2024  10:13:47 -0500 (0:00:00.065)       0:06:21.439 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Wednesday 11 December 2024  10:13:47 -0500 (0:00:00.070)       0:06:21.510 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Wednesday 11 December 2024  10:13:47 -0500 (0:00:00.069)       0:06:21.580 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Wednesday 11 December 2024  10:13:47 -0500 (0:00:00.069)       0:06:21.649 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Wednesday 11 December 2024  10:13:47 -0500 (0:00:00.053)       0:06:21.703 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Wednesday 11 December 2024  10:13:47 -0500 (0:00:00.052)       0:06:21.755 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Wednesday 11 December 2024  10:13:47 -0500 (0:00:00.059)       0:06:21.814 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Wednesday 11 December 2024  10:13:47 -0500 (0:00:00.049)       0:06:21.864 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Wednesday 11 December 2024  10:13:47 -0500 (0:00:00.045)       0:06:21.909 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Wednesday 11 December 2024  10:13:47 -0500 (0:00:00.044)       0:06:21.954 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Wednesday 11 December 2024  10:13:47 -0500 (0:00:00.044)       0:06:21.998 **** 
ok: [managed-node3] => {
    "bytes": 9663676416,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Wednesday 11 December 2024  10:13:48 -0500 (0:00:00.495)       0:06:22.493 **** 
ok: [managed-node3] => {
    "bytes": 9663676416,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Wednesday 11 December 2024  10:13:48 -0500 (0:00:00.414)       0:06:22.907 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_expected_size": "9663676416"
    },
    "changed": false
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Wednesday 11 December 2024  10:13:48 -0500 (0:00:00.063)       0:06:22.971 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "9663676416"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Wednesday 11 December 2024  10:13:48 -0500 (0:00:00.047)       0:06:23.019 **** 
ok: [managed-node3] => {
    "bytes": 10726680821,
    "changed": false,
    "lvm": "9g",
    "parted": "9GiB",
    "size": "9 GiB"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Wednesday 11 December 2024  10:13:49 -0500 (0:00:00.409)       0:06:23.429 **** 
skipping: [managed-node3] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Wednesday 11 December 2024  10:13:49 -0500 (0:00:00.068)       0:06:23.497 **** 
skipping: [managed-node3] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Wednesday 11 December 2024  10:13:49 -0500 (0:00:00.054)       0:06:23.551 **** 
skipping: [managed-node3] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Wednesday 11 December 2024  10:13:49 -0500 (0:00:00.059)       0:06:23.611 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Wednesday 11 December 2024  10:13:49 -0500 (0:00:00.052)       0:06:23.664 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Wednesday 11 December 2024  10:13:49 -0500 (0:00:00.043)       0:06:23.707 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Wednesday 11 December 2024  10:13:49 -0500 (0:00:00.044)       0:06:23.751 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Wednesday 11 December 2024  10:13:49 -0500 (0:00:00.044)       0:06:23.796 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Wednesday 11 December 2024  10:13:49 -0500 (0:00:00.122)       0:06:23.919 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Wednesday 11 December 2024  10:13:49 -0500 (0:00:00.050)       0:06:23.969 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Wednesday 11 December 2024  10:13:49 -0500 (0:00:00.051)       0:06:24.021 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Wednesday 11 December 2024  10:13:49 -0500 (0:00:00.058)       0:06:24.079 **** 
skipping: [managed-node3] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Wednesday 11 December 2024  10:13:49 -0500 (0:00:00.049)       0:06:24.129 **** 
skipping: [managed-node3] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Wednesday 11 December 2024  10:13:49 -0500 (0:00:00.042)       0:06:24.171 **** 
skipping: [managed-node3] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Wednesday 11 December 2024  10:13:49 -0500 (0:00:00.043)       0:06:24.214 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Wednesday 11 December 2024  10:13:49 -0500 (0:00:00.044)       0:06:24.259 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Wednesday 11 December 2024  10:13:50 -0500 (0:00:00.043)       0:06:24.302 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Wednesday 11 December 2024  10:13:50 -0500 (0:00:00.042)       0:06:24.345 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Wednesday 11 December 2024  10:13:50 -0500 (0:00:00.043)       0:06:24.388 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Wednesday 11 December 2024  10:13:50 -0500 (0:00:00.054)       0:06:24.442 **** 
ok: [managed-node3] => {
    "storage_test_actual_size": {
        "bytes": 9663676416,
        "changed": false,
        "failed": false,
        "lvm": "9g",
        "parted": "9GiB",
        "size": "9 GiB"
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Wednesday 11 December 2024  10:13:50 -0500 (0:00:00.046)       0:06:24.489 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "9663676416"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Wednesday 11 December 2024  10:13:50 -0500 (0:00:00.044)       0:06:24.534 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Wednesday 11 December 2024  10:13:50 -0500 (0:00:00.052)       0:06:24.587 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.024585",
    "end": "2024-12-11 10:13:50.666969",
    "rc": 0,
    "start": "2024-12-11 10:13:50.642384"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Wednesday 11 December 2024  10:13:50 -0500 (0:00:00.449)       0:06:25.036 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Wednesday 11 December 2024  10:13:50 -0500 (0:00:00.049)       0:06:25.085 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Wednesday 11 December 2024  10:13:50 -0500 (0:00:00.054)       0:06:25.140 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Wednesday 11 December 2024  10:13:50 -0500 (0:00:00.057)       0:06:25.197 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Wednesday 11 December 2024  10:13:50 -0500 (0:00:00.053)       0:06:25.251 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Wednesday 11 December 2024  10:13:51 -0500 (0:00:00.054)       0:06:25.305 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Wednesday 11 December 2024  10:13:51 -0500 (0:00:00.054)       0:06:25.359 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Wednesday 11 December 2024  10:13:51 -0500 (0:00:00.049)       0:06:25.409 **** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Wednesday 11 December 2024  10:13:51 -0500 (0:00:00.048)       0:06:25.457 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Test for correct handling of offline resize in safe mode] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:377
Wednesday 11 December 2024  10:13:51 -0500 (0:00:00.046)       0:06:25.503 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed-node3

TASK [Store global variable value copy] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4
Wednesday 11 December 2024  10:13:51 -0500 (0:00:00.081)       0:06:25.584 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_pools_global": [],
        "storage_safe_mode_global": false,
        "storage_volumes_global": []
    },
    "changed": false
}

TASK [Verify role raises correct error] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10
Wednesday 11 December 2024  10:13:51 -0500 (0:00:00.066)       0:06:25.650 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:13:51 -0500 (0:00:00.148)       0:06:25.798 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:13:51 -0500 (0:00:00.077)       0:06:25.876 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:13:51 -0500 (0:00:00.053)       0:06:25.930 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:13:51 -0500 (0:00:00.109)       0:06:26.039 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:13:51 -0500 (0:00:00.043)       0:06:26.083 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:13:51 -0500 (0:00:00.042)       0:06:26.126 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:13:51 -0500 (0:00:00.046)       0:06:26.172 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:13:51 -0500 (0:00:00.044)       0:06:26.217 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:13:52 -0500 (0:00:00.111)       0:06:26.329 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:13:52 -0500 (0:00:00.069)       0:06:26.398 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "5g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:13:52 -0500 (0:00:00.089)       0:06:26.488 **** 
ok: [managed-node3] => {
    "storage_volumes": []
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:13:52 -0500 (0:00:00.073)       0:06:26.562 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:13:52 -0500 (0:00:00.113)       0:06:26.676 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:13:52 -0500 (0:00:00.053)       0:06:26.729 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:13:52 -0500 (0:00:00.047)       0:06:26.776 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:13:52 -0500 (0:00:00.043)       0:06:26.820 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:13:52 -0500 (0:00:00.065)       0:06:26.885 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:13:52 -0500 (0:00:00.043)       0:06:26.928 **** 
fatal: [managed-node3]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

device 'foo-test1' must be unmounted to be resized in safe mode

TASK [fedora.linux_system_roles.storage : Failed message] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109
Wednesday 11 December 2024  10:13:57 -0500 (0:00:04.774)       0:06:31.703 **** 
fatal: [managed-node3]: FAILED! => {
    "changed": false
}

MSG:

{'msg': "device 'foo-test1' must be unmounted to be resized in safe mode", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'encryption_clevis_pin': None, 'encryption_tang_url': None, 'encryption_tang_thumbprint': None, 'grow_to_fill': False, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'shared': False, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '5g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}]}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'grow_to_fill': False, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, 'shared': False}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:13:57 -0500 (0:00:00.085)       0:06:31.789 **** 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23
Wednesday 11 December 2024  10:13:57 -0500 (0:00:00.043)       0:06:31.833 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the blivet output and error message are correct] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28
Wednesday 11 December 2024  10:13:57 -0500 (0:00:00.056)       0:06:31.889 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify correct exception or error message] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39
Wednesday 11 December 2024  10:13:57 -0500 (0:00:00.062)       0:06:31.952 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up] ****************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:394
Wednesday 11 December 2024  10:13:57 -0500 (0:00:00.053)       0:06:32.005 **** 

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Wednesday 11 December 2024  10:13:57 -0500 (0:00:00.084)       0:06:32.090 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Wednesday 11 December 2024  10:13:57 -0500 (0:00:00.073)       0:06:32.163 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Wednesday 11 December 2024  10:13:57 -0500 (0:00:00.107)       0:06:32.270 **** 
skipping: [managed-node3] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node3] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS.yml",
    "skip_reason": "Conditional result was False"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}
ok: [managed-node3] => (item=CentOS_8.yml) => {
    "ansible_facts": {
        "blivet_package_list": [
            "python3-blivet",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap",
            "vdo",
            "kmod-kvdo",
            "xfsprogs",
            "stratisd",
            "stratis-cli",
            "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_8.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "CentOS_8.yml"
}

TASK [fedora.linux_system_roles.storage : Check if system is ostree] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25
Wednesday 11 December 2024  10:13:58 -0500 (0:00:00.109)       0:06:32.379 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30
Wednesday 11 December 2024  10:13:58 -0500 (0:00:00.043)       0:06:32.422 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Wednesday 11 December 2024  10:13:58 -0500 (0:00:00.044)       0:06:32.467 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Wednesday 11 December 2024  10:13:58 -0500 (0:00:00.045)       0:06:32.512 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Wednesday 11 December 2024  10:13:58 -0500 (0:00:00.046)       0:06:32.559 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3

TASK [fedora.linux_system_roles.storage : Make sure blivet is available] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Wednesday 11 December 2024  10:13:58 -0500 (0:00:00.086)       0:06:32.645 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Show storage_pools] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9
Wednesday 11 December 2024  10:13:58 -0500 (0:00:00.047)       0:06:32.693 **** 
ok: [managed-node3] => {
    "storage_pools": [
        {
            "disks": [
                "sda"
            ],
            "name": "foo",
            "state": "absent",
            "volumes": [
                {
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "5g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Show storage_volumes] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14
Wednesday 11 December 2024  10:13:58 -0500 (0:00:00.057)       0:06:32.750 **** 
ok: [managed-node3] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : Get required packages] ***************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19
Wednesday 11 December 2024  10:13:58 -0500 (0:00:00.046)       0:06:32.797 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31
Wednesday 11 December 2024  10:13:58 -0500 (0:00:00.043)       0:06:32.840 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37
Wednesday 11 December 2024  10:13:58 -0500 (0:00:00.044)       0:06:32.885 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Get service facts] *******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51
Wednesday 11 December 2024  10:13:58 -0500 (0:00:00.044)       0:06:32.929 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57
Wednesday 11 December 2024  10:13:58 -0500 (0:00:00.105)       0:06:33.035 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63
Wednesday 11 December 2024  10:13:58 -0500 (0:00:00.061)       0:06:33.097 **** 

TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69
Wednesday 11 December 2024  10:13:58 -0500 (0:00:00.039)       0:06:33.136 **** 
changed: [managed-node3] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/sda",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/sdd",
        "/dev/sde",
        "/dev/sdf",
        "/dev/sdg",
        "/dev/sdh",
        "/dev/sdi",
        "/dev/xvda1"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "absent",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83
Wednesday 11 December 2024  10:14:03 -0500 (0:00:04.908)       0:06:38.045 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Wednesday 11 December 2024  10:14:03 -0500 (0:00:00.069)       0:06:38.114 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733929981.6155782,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 8,
        "charset": "us-ascii",
        "checksum": "306e713ef45e6db45ca0244c07767212954e6abb",
        "ctime": 1733929981.6125782,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 174063825,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "text/plain",
        "mode": "0644",
        "mtime": 1733929981.6125782,
        "nlink": 1,
        "path": "/etc/fstab",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": true,
        "rusr": true,
        "size": 1394,
        "uid": 0,
        "version": "1265219349",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95
Wednesday 11 December 2024  10:14:04 -0500 (0:00:00.459)       0:06:38.573 **** 
ok: [managed-node3] => {
    "backup": "",
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113
Wednesday 11 December 2024  10:14:04 -0500 (0:00:00.450)       0:06:39.024 **** 

TASK [fedora.linux_system_roles.storage : Show blivet_output] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119
Wednesday 11 December 2024  10:14:04 -0500 (0:00:00.038)       0:06:39.062 **** 
ok: [managed-node3] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/sda",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/sdd",
            "/dev/sde",
            "/dev/sdf",
            "/dev/sdg",
            "/dev/sdh",
            "/dev/sdi",
            "/dev/xvda1"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "absent",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128
Wednesday 11 December 2024  10:14:04 -0500 (0:00:00.050)       0:06:39.113 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "sda"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_clevis_pin": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "encryption_tang_thumbprint": null,
                "encryption_tang_url": null,
                "grow_to_fill": false,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "shared": false,
                "state": "absent",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [
                            "sda"
                        ],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_group": null,
                        "mount_mode": null,
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "mount_user": null,
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "raid_stripe_size": null,
                        "size": "5g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Wednesday 11 December 2024  10:14:04 -0500 (0:00:00.048)       0:06:39.162 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] **************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Wednesday 11 December 2024  10:14:04 -0500 (0:00:00.047)       0:06:39.209 **** 
changed: [managed-node3] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159
Wednesday 11 December 2024  10:14:05 -0500 (0:00:00.419)       0:06:39.629 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Set up new/current mounts] ***********
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164
Wednesday 11 December 2024  10:14:06 -0500 (0:00:00.668)       0:06:40.297 **** 

TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175
Wednesday 11 December 2024  10:14:06 -0500 (0:00:00.039)       0:06:40.337 **** 

TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187
Wednesday 11 December 2024  10:14:06 -0500 (0:00:00.040)       0:06:40.377 **** 
ok: [managed-node3] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195
Wednesday 11 December 2024  10:14:06 -0500 (0:00:00.660)       0:06:41.038 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "atime": 1733928636.989118,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1716968941.893,
        "dev": 51713,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 135,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1716968586.525,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "1157759751",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200
Wednesday 11 December 2024  10:14:07 -0500 (0:00:00.424)       0:06:41.462 **** 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222
Wednesday 11 December 2024  10:14:07 -0500 (0:00:00.069)       0:06:41.532 **** 
ok: [managed-node3]

TASK [Verify role results] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:407
Wednesday 11 December 2024  10:14:08 -0500 (0:00:00.787)       0:06:42.320 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed-node3

TASK [Print out pool information] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2
Wednesday 11 December 2024  10:14:08 -0500 (0:00:00.079)       0:06:42.399 **** 
ok: [managed-node3] => {
    "_storage_pools_list": [
        {
            "disks": [
                "sda"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_clevis_pin": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "encryption_tang_thumbprint": null,
            "encryption_tang_url": null,
            "grow_to_fill": false,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "shared": false,
            "state": "absent",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [
                        "sda"
                    ],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_group": null,
                    "mount_mode": null,
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "mount_user": null,
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "raid_stripe_size": null,
                    "size": "5g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7
Wednesday 11 December 2024  10:14:08 -0500 (0:00:00.049)       0:06:42.449 **** 
skipping: [managed-node3] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15
Wednesday 11 December 2024  10:14:08 -0500 (0:00:00.044)       0:06:42.493 **** 
ok: [managed-node3] => {
    "changed": false,
    "info": {
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdd": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdd",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sde": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sde",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdf": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdf",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdg": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdg",
            "size": "1T",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdh": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdh",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdi": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/sdi",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda": {
            "fstype": "",
            "label": "",
            "mountpoint": "",
            "name": "/dev/xvda",
            "size": "250G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/xvda1": {
            "fstype": "xfs",
            "label": "",
            "mountpoint": "/",
            "name": "/dev/xvda1",
            "size": "250G",
            "type": "partition",
            "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20
Wednesday 11 December 2024  10:14:08 -0500 (0:00:00.427)       0:06:42.921 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.002766",
    "end": "2024-12-11 10:14:08.987114",
    "rc": 0,
    "start": "2024-12-11 10:14:08.984348"
}

STDOUT:


# system_role:storage
#
# /etc/fstab
# Created by anaconda on Wed May 29 07:43:06 2024
#
# Accessible filesystems, by reference, are maintained under '/dev/disk/'.
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.
#
# After editing this file, run 'systemctl daemon-reload' to update systemd
# units generated from this file.
#
UUID=fe591198-9082-4b15-9b62-e83518524cd2 /                       xfs     defaults        0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0
ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25
Wednesday 11 December 2024  10:14:09 -0500 (0:00:00.514)       0:06:43.435 **** 
ok: [managed-node3] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.002823",
    "end": "2024-12-11 10:14:09.468922",
    "failed_when_result": false,
    "rc": 0,
    "start": "2024-12-11 10:14:09.466099"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34
Wednesday 11 December 2024  10:14:09 -0500 (0:00:00.420)       0:06:43.855 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed-node3

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5
Wednesday 11 December 2024  10:14:09 -0500 (0:00:00.112)       0:06:43.968 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [Get VG shared value status] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18
Wednesday 11 December 2024  10:14:09 -0500 (0:00:00.089)       0:06:44.057 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that VG shared value checks out] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24
Wednesday 11 December 2024  10:14:09 -0500 (0:00:00.063)       0:06:44.121 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify pool subset] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34
Wednesday 11 December 2024  10:14:09 -0500 (0:00:00.052)       0:06:44.173 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2
Wednesday 11 December 2024  10:14:09 -0500 (0:00:00.093)       0:06:44.267 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "0",
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8
Wednesday 11 December 2024  10:14:10 -0500 (0:00:00.057)       0:06:44.325 **** 

TASK [Set pvs lvm length] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17
Wednesday 11 December 2024  10:14:10 -0500 (0:00:00.057)       0:06:44.382 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": "0"
    },
    "changed": false
}

TASK [Set pool pvs] ************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22
Wednesday 11 December 2024  10:14:10 -0500 (0:00:00.073)       0:06:44.455 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": []
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27
Wednesday 11 December 2024  10:14:10 -0500 (0:00:00.069)       0:06:44.525 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36
Wednesday 11 December 2024  10:14:10 -0500 (0:00:00.145)       0:06:44.671 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41
Wednesday 11 December 2024  10:14:10 -0500 (0:00:00.054)       0:06:44.725 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [Set expected pv type] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46
Wednesday 11 December 2024  10:14:10 -0500 (0:00:00.048)       0:06:44.774 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51
Wednesday 11 December 2024  10:14:10 -0500 (0:00:00.046)       0:06:44.821 **** 

TASK [Check that blivet supports PV grow to fill] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64
Wednesday 11 December 2024  10:14:10 -0500 (0:00:00.039)       0:06:44.860 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0
}

STDOUT:

False



STDERR:

Shared connection to 10.31.41.57 closed.


TASK [Verify that PVs fill the whole devices when they should] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73
Wednesday 11 December 2024  10:14:11 -0500 (0:00:00.493)       0:06:45.354 **** 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83
Wednesday 11 December 2024  10:14:11 -0500 (0:00:00.038)       0:06:45.392 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed-node3

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8
Wednesday 11 December 2024  10:14:11 -0500 (0:00:00.098)       0:06:45.490 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14
Wednesday 11 December 2024  10:14:11 -0500 (0:00:00.075)       0:06:45.565 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19
Wednesday 11 December 2024  10:14:11 -0500 (0:00:00.064)       0:06:45.630 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24
Wednesday 11 December 2024  10:14:11 -0500 (0:00:00.063)       0:06:45.693 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md chunk size regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29
Wednesday 11 December 2024  10:14:11 -0500 (0:00:00.059)       0:06:45.753 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37
Wednesday 11 December 2024  10:14:11 -0500 (0:00:00.053)       0:06:45.807 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46
Wednesday 11 December 2024  10:14:11 -0500 (0:00:00.100)       0:06:45.907 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55
Wednesday 11 December 2024  10:14:11 -0500 (0:00:00.043)       0:06:45.950 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64
Wednesday 11 December 2024  10:14:11 -0500 (0:00:00.044)       0:06:45.995 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74
Wednesday 11 December 2024  10:14:11 -0500 (0:00:00.041)       0:06:46.037 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variables used by tests] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83
Wednesday 11 December 2024  10:14:11 -0500 (0:00:00.044)       0:06:46.081 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_chunk_size_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86
Wednesday 11 December 2024  10:14:11 -0500 (0:00:00.055)       0:06:46.136 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed-node3

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2
Wednesday 11 December 2024  10:14:11 -0500 (0:00:00.089)       0:06:46.226 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed-node3

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8
Wednesday 11 December 2024  10:14:12 -0500 (0:00:00.108)       0:06:46.335 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16
Wednesday 11 December 2024  10:14:12 -0500 (0:00:00.058)       0:06:46.393 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20
Wednesday 11 December 2024  10:14:12 -0500 (0:00:00.081)       0:06:46.475 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV stripe size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27
Wednesday 11 December 2024  10:14:12 -0500 (0:00:00.068)       0:06:46.543 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested stripe size] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31
Wednesday 11 December 2024  10:14:12 -0500 (0:00:00.068)       0:06:46.612 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected stripe size] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37
Wednesday 11 December 2024  10:14:12 -0500 (0:00:00.141)       0:06:46.753 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check stripe size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42
Wednesday 11 December 2024  10:14:12 -0500 (0:00:00.077)       0:06:46.831 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89
Wednesday 11 December 2024  10:14:12 -0500 (0:00:00.058)       0:06:46.889 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed-node3

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2
Wednesday 11 December 2024  10:14:12 -0500 (0:00:00.111)       0:06:47.000 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed-node3

TASK [Get information about thinpool] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8
Wednesday 11 December 2024  10:14:12 -0500 (0:00:00.085)       0:06:47.086 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16
Wednesday 11 December 2024  10:14:12 -0500 (0:00:00.055)       0:06:47.141 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22
Wednesday 11 December 2024  10:14:12 -0500 (0:00:00.074)       0:06:47.216 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26
Wednesday 11 December 2024  10:14:12 -0500 (0:00:00.053)       0:06:47.269 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_thin_status": null
    },
    "changed": false
}

TASK [Check member encryption] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92
Wednesday 11 December 2024  10:14:13 -0500 (0:00:00.055)       0:06:47.325 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed-node3

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5
Wednesday 11 December 2024  10:14:13 -0500 (0:00:00.089)       0:06:47.415 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10
Wednesday 11 December 2024  10:14:13 -0500 (0:00:00.052)       0:06:47.467 **** 

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17
Wednesday 11 December 2024  10:14:13 -0500 (0:00:00.123)       0:06:47.591 **** 

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24
Wednesday 11 December 2024  10:14:13 -0500 (0:00:00.057)       0:06:47.648 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95
Wednesday 11 December 2024  10:14:13 -0500 (0:00:00.061)       0:06:47.709 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed-node3

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2
Wednesday 11 December 2024  10:14:13 -0500 (0:00:00.112)       0:06:47.822 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed-node3

TASK [Get information about VDO deduplication] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8
Wednesday 11 December 2024  10:14:13 -0500 (0:00:00.080)       0:06:47.903 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15
Wednesday 11 December 2024  10:14:13 -0500 (0:00:00.046)       0:06:47.949 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21
Wednesday 11 December 2024  10:14:13 -0500 (0:00:00.061)       0:06:48.010 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about VDO compression] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27
Wednesday 11 December 2024  10:14:13 -0500 (0:00:00.091)       0:06:48.102 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is off] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34
Wednesday 11 December 2024  10:14:13 -0500 (0:00:00.077)       0:06:48.180 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check if VDO deduplication is on] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40
Wednesday 11 December 2024  10:14:13 -0500 (0:00:00.079)       0:06:48.259 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46
Wednesday 11 December 2024  10:14:14 -0500 (0:00:00.053)       0:06:48.313 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [Check Stratis] ***********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98
Wednesday 11 December 2024  10:14:14 -0500 (0:00:00.054)       0:06:48.367 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed-node3

TASK [Run 'stratis report'] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6
Wednesday 11 December 2024  10:14:14 -0500 (0:00:00.170)       0:06:48.538 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about Stratis] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11
Wednesday 11 December 2024  10:14:14 -0500 (0:00:00.056)       0:06:48.595 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the pools was created] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15
Wednesday 11 December 2024  10:14:14 -0500 (0:00:00.060)       0:06:48.655 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that encryption is correctly set] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25
Wednesday 11 December 2024  10:14:14 -0500 (0:00:00.066)       0:06:48.722 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that Clevis/Tang encryption is correctly set] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34
Wednesday 11 December 2024  10:14:14 -0500 (0:00:00.079)       0:06:48.809 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Reset variable used by test] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44
Wednesday 11 December 2024  10:14:14 -0500 (0:00:00.069)       0:06:48.878 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_stratis_report": null
    },
    "changed": false
}

TASK [Clean up test variables] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101
Wednesday 11 December 2024  10:14:14 -0500 (0:00:00.064)       0:06:48.943 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Verify the volumes] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3
Wednesday 11 December 2024  10:14:14 -0500 (0:00:00.053)       0:06:48.996 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed-node3

TASK [Set storage volume test variables] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2
Wednesday 11 December 2024  10:14:14 -0500 (0:00:00.090)       0:06:49.087 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": false,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [Run test verify for {{ storage_test_volume_subset }}] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19
Wednesday 11 December 2024  10:14:14 -0500 (0:00:00.054)       0:06:49.141 **** 
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed-node3
included: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed-node3

TASK [Get expected mount device based on device type] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7
Wednesday 11 December 2024  10:14:15 -0500 (0:00:00.197)       0:06:49.339 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11
Wednesday 11 December 2024  10:14:15 -0500 (0:00:00.190)       0:06:49.529 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_mount_expected_mount_point": "/opt/test1",
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Get information about the mountpoint directory] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19
Wednesday 11 December 2024  10:14:15 -0500 (0:00:00.081)       0:06:49.611 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28
Wednesday 11 December 2024  10:14:15 -0500 (0:00:00.087)       0:06:49.698 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory user] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36
Wednesday 11 December 2024  10:14:15 -0500 (0:00:00.079)       0:06:49.777 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory group] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42
Wednesday 11 December 2024  10:14:15 -0500 (0:00:00.068)       0:06:49.846 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify mount directory permissions] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48
Wednesday 11 December 2024  10:14:15 -0500 (0:00:00.065)       0:06:49.911 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get path of test volume device] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57
Wednesday 11 December 2024  10:14:15 -0500 (0:00:00.060)       0:06:49.971 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63
Wednesday 11 December 2024  10:14:15 -0500 (0:00:00.044)       0:06:50.016 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69
Wednesday 11 December 2024  10:14:15 -0500 (0:00:00.044)       0:06:50.060 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79
Wednesday 11 December 2024  10:14:15 -0500 (0:00:00.044)       0:06:50.105 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_found_mount_stat": null,
        "storage_test_mount_expected_mount_point": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2
Wednesday 11 December 2024  10:14:15 -0500 (0:00:00.046)       0:06:50.152 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "0",
        "storage_test_fstab_expected_mount_options_matches": "0",
        "storage_test_fstab_expected_mount_point_matches": "0",
        "storage_test_fstab_id_matches": [],
        "storage_test_fstab_mount_options_matches": [],
        "storage_test_fstab_mount_point_matches": []
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17
Wednesday 11 December 2024  10:14:15 -0500 (0:00:00.093)       0:06:50.245 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24
Wednesday 11 December 2024  10:14:16 -0500 (0:00:00.087)       0:06:50.332 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33
Wednesday 11 December 2024  10:14:16 -0500 (0:00:00.084)       0:06:50.416 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fingerprint] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45
Wednesday 11 December 2024  10:14:16 -0500 (0:00:00.148)       0:06:50.565 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:52
Wednesday 11 December 2024  10:14:16 -0500 (0:00:00.073)       0:06:50.638 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6
Wednesday 11 December 2024  10:14:16 -0500 (0:00:00.070)       0:06:50.709 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify fs label] *********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14
Wednesday 11 December 2024  10:14:16 -0500 (0:00:00.081)       0:06:50.791 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [See whether the device node is present] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3
Wednesday 11 December 2024  10:14:16 -0500 (0:00:00.077)       0:06:50.868 **** 
ok: [managed-node3] => {
    "changed": false,
    "stat": {
        "exists": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9
Wednesday 11 December 2024  10:14:17 -0500 (0:00:00.545)       0:06:51.413 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16
Wednesday 11 December 2024  10:14:17 -0500 (0:00:00.085)       0:06:51.499 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23
Wednesday 11 December 2024  10:14:17 -0500 (0:00:00.127)       0:06:51.626 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Process volume type (set initial value) (1/2)] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29
Wednesday 11 December 2024  10:14:17 -0500 (0:00:00.066)       0:06:51.693 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [Process volume type (get RAID value) (2/2)] ******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33
Wednesday 11 December 2024  10:14:17 -0500 (0:00:00.106)       0:06:51.799 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38
Wednesday 11 December 2024  10:14:17 -0500 (0:00:00.155)       0:06:51.955 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3
Wednesday 11 December 2024  10:14:17 -0500 (0:00:00.103)       0:06:52.058 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10
Wednesday 11 December 2024  10:14:17 -0500 (0:00:00.079)       0:06:52.138 **** 
ok: [managed-node3] => {
    "changed": false,
    "rc": 0,
    "results": []
}

MSG:

Nothing to do

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16
Wednesday 11 December 2024  10:14:20 -0500 (0:00:03.106)       0:06:55.244 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22
Wednesday 11 December 2024  10:14:20 -0500 (0:00:00.045)       0:06:55.290 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.043)       0:06:55.333 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.043)       0:06:55.377 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.046)       0:06:55.423 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.047)       0:06:55.471 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.043)       0:06:55.514 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.042)       0:06:55.557 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set test variables] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.056)       0:06:55.614 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.111)       0:06:55.725 **** 
ok: [managed-node3] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.049)       0:06:55.775 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.044)       0:06:55.820 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.044)       0:06:55.864 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clear test variables] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.045)       0:06:55.910 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [Get information about RAID] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.045)       0:06:55.956 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set active devices regex] ************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.045)       0:06:56.001 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set spare devices regex] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.044)       0:06:56.046 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set md version regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.045)       0:06:56.091 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set chunk size regex] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.042)       0:06:56.134 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the chunk size] ****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.044)       0:06:56.179 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID active devices count] *****************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.043)       0:06:56.223 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID spare devices count] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54
Wednesday 11 December 2024  10:14:21 -0500 (0:00:00.044)       0:06:56.267 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID metadata version] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62
Wednesday 11 December 2024  10:14:22 -0500 (0:00:00.043)       0:06:56.311 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check RAID chunk size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70
Wednesday 11 December 2024  10:14:22 -0500 (0:00:00.043)       0:06:56.354 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the actual size of the volume] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3
Wednesday 11 December 2024  10:14:22 -0500 (0:00:00.125)       0:06:56.479 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested size of the volume] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11
Wednesday 11 December 2024  10:14:22 -0500 (0:00:00.070)       0:06:56.549 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20
Wednesday 11 December 2024  10:14:22 -0500 (0:00:00.075)       0:06:56.625 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28
Wednesday 11 December 2024  10:14:22 -0500 (0:00:00.075)       0:06:56.700 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "9663676416"
}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32
Wednesday 11 December 2024  10:14:22 -0500 (0:00:00.071)       0:06:56.771 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show test pool] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46
Wednesday 11 December 2024  10:14:22 -0500 (0:00:00.067)       0:06:56.839 **** 
skipping: [managed-node3] => {}

TASK [Show test blockinfo] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50
Wednesday 11 December 2024  10:14:22 -0500 (0:00:00.069)       0:06:56.909 **** 
skipping: [managed-node3] => {}

TASK [Show test pool size] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54
Wednesday 11 December 2024  10:14:22 -0500 (0:00:00.067)       0:06:56.977 **** 
skipping: [managed-node3] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58
Wednesday 11 December 2024  10:14:22 -0500 (0:00:00.072)       0:06:57.050 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default thin pool reserved space values] *********************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67
Wednesday 11 December 2024  10:14:22 -0500 (0:00:00.064)       0:06:57.115 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default minimal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71
Wednesday 11 December 2024  10:14:22 -0500 (0:00:00.068)       0:06:57.183 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Default maximal thin pool reserved space size] ***************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76
Wednesday 11 December 2024  10:14:22 -0500 (0:00:00.072)       0:06:57.255 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate maximum usable space in thin pool] *****************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82
Wednesday 11 December 2024  10:14:23 -0500 (0:00:00.076)       0:06:57.332 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply upper size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86
Wednesday 11 December 2024  10:14:23 -0500 (0:00:00.081)       0:06:57.414 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Apply lower size limit to max usable thin pool space] ********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91
Wednesday 11 December 2024  10:14:23 -0500 (0:00:00.086)       0:06:57.501 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Convert maximum usable thin pool space from int to Size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96
Wednesday 11 December 2024  10:14:23 -0500 (0:00:00.230)       0:06:57.731 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show max thin pool size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101
Wednesday 11 December 2024  10:14:23 -0500 (0:00:00.083)       0:06:57.815 **** 
skipping: [managed-node3] => {}

TASK [Show volume thin pool size] **********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105
Wednesday 11 December 2024  10:14:23 -0500 (0:00:00.142)       0:06:57.958 **** 
skipping: [managed-node3] => {}

TASK [Show test volume size] ***************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109
Wednesday 11 December 2024  10:14:23 -0500 (0:00:00.124)       0:06:58.082 **** 
skipping: [managed-node3] => {}

TASK [Establish base value for expected thin pool size] ************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113
Wednesday 11 December 2024  10:14:23 -0500 (0:00:00.084)       0:06:58.166 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120
Wednesday 11 December 2024  10:14:23 -0500 (0:00:00.083)       0:06:58.250 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected thin pool volume size] *****************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127
Wednesday 11 December 2024  10:14:24 -0500 (0:00:00.089)       0:06:58.340 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Calculate the expected thin pool volume size based on percentage value] ***
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131
Wednesday 11 December 2024  10:14:24 -0500 (0:00:00.091)       0:06:58.431 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Replace expected volume size with calculated value] **********************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137
Wednesday 11 December 2024  10:14:24 -0500 (0:00:00.068)       0:06:58.500 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Show actual size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143
Wednesday 11 December 2024  10:14:24 -0500 (0:00:00.089)       0:06:58.589 **** 
ok: [managed-node3] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [Show expected size] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147
Wednesday 11 December 2024  10:14:24 -0500 (0:00:00.078)       0:06:58.668 **** 
ok: [managed-node3] => {
    "storage_test_expected_size": "9663676416"
}

TASK [Assert expected size is actual size] *************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151
Wednesday 11 December 2024  10:14:24 -0500 (0:00:00.071)       0:06:58.740 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5
Wednesday 11 December 2024  10:14:24 -0500 (0:00:00.088)       0:06:58.828 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV segment type] *****************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13
Wednesday 11 December 2024  10:14:24 -0500 (0:00:00.069)       0:06:58.897 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check segment type] ******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17
Wednesday 11 December 2024  10:14:24 -0500 (0:00:00.078)       0:06:58.976 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set LV cache size] *******************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24
Wednesday 11 December 2024  10:14:24 -0500 (0:00:00.152)       0:06:59.129 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Parse the requested cache size] ******************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31
Wednesday 11 December 2024  10:14:24 -0500 (0:00:00.099)       0:06:59.228 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Set expected cache size] *************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37
Wednesday 11 December 2024  10:14:25 -0500 (0:00:00.077)       0:06:59.305 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42
Wednesday 11 December 2024  10:14:25 -0500 (0:00:00.068)       0:06:59.373 **** 
skipping: [managed-node3] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25
Wednesday 11 December 2024  10:14:25 -0500 (0:00:00.061)       0:06:59.434 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44
Wednesday 11 December 2024  10:14:25 -0500 (0:00:00.058)       0:06:59.493 **** 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54
Wednesday 11 December 2024  10:14:25 -0500 (0:00:00.038)       0:06:59.532 **** 
ok: [managed-node3] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}
META: ran handlers
META: ran handlers

PLAY RECAP *********************************************************************
managed-node3              : ok=1637 changed=27   unreachable=0    failed=4    skipped=1840 rescued=4    ignored=0   

Wednesday 11 December 2024  10:14:25 -0500 (0:00:00.021)       0:06:59.553 **** 
=============================================================================== 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 6.71s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 5.91s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 5.74s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 5.08s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.96s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.91s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.89s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.83s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.78s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.77s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.70s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.59s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.58s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.57s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.50s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.47s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.36s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.34s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.22s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 
fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.12s
/tmp/collections-zNM/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69