ansible-playbook [core 2.12.6]
  config file = /etc/ansible/ansible.cfg
  configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
  ansible python module location = /usr/lib/python3.9/site-packages/ansible
  ansible collection location = /tmp/tmp5bkr4li_
  executable location = /usr/bin/ansible-playbook
  python version = 3.9.13 (main, May 18 2022, 00:00:00) [GCC 11.3.1 20220421 (Red Hat 11.3.1-2)]
  jinja version = 2.11.3
  libyaml = True
Using /etc/ansible/ansible.cfg as config file
Skipping callback 'debug', as we already have a stdout callback.
Skipping callback 'default', as we already have a stdout callback.
Skipping callback 'minimal', as we already have a stdout callback.
Skipping callback 'oneline', as we already have a stdout callback.

PLAYBOOK: rhel-7_setup.yml *****************************************************
1 plays in /cache/rhel-7_setup.yml

PLAY [Setup repos] *************************************************************
META: ran handlers

TASK [set up internal repositories] ********************************************
task path: /cache/rhel-7_setup.yml:5
Thursday 21 July 2022  15:39:19 +0000 (0:00:00.018)       0:00:00.018 ********* 
changed: [/cache/rhel-7.qcow2] => (item=None) => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
changed: [/cache/rhel-7.qcow2] => (item=None) => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
changed: [/cache/rhel-7.qcow2] => (item=None) => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
changed: [/cache/rhel-7.qcow2] => (item=None) => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
changed: [/cache/rhel-7.qcow2] => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
META: ran handlers
META: ran handlers

PLAY RECAP *********************************************************************
/cache/rhel-7.qcow2        : ok=1    changed=1    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   

Thursday 21 July 2022  15:39:21 +0000 (0:00:01.419)       0:00:01.438 ********* 
=============================================================================== 
set up internal repositories -------------------------------------------- 1.42s
/cache/rhel-7_setup.yml:5 -----------------------------------------------------

PLAYBOOK: tests_misc.yml *******************************************************
1 plays in /tmp/tmpaxjje44y/tests/tests_misc.yml

PLAY [all] *********************************************************************

TASK [Gathering Facts] *********************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:2
Thursday 21 July 2022  15:39:21 +0000 (0:00:00.020)       0:00:01.459 ********* 
ok: [/cache/rhel-7.qcow2]
META: ran handlers

TASK [include_role : linux-system-roles.storage] *******************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:16
Thursday 21 July 2022  15:39:22 +0000 (0:00:01.001)       0:00:02.460 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:39:22 +0000 (0:00:00.032)       0:00:02.493 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:39:22 +0000 (0:00:00.029)       0:00:02.523 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:39:22 +0000 (0:00:00.410)       0:00:02.933 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:39:22 +0000 (0:00:00.056)       0:00:02.989 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:39:22 +0000 (0:00:00.028)       0:00:03.018 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:39:22 +0000 (0:00:00.030)       0:00:03.048 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:39:22 +0000 (0:00:00.045)       0:00:03.094 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:39:22 +0000 (0:00:00.018)       0:00:03.112 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "changed": true,
    "changes": {
        "installed": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "rc": 0,
    "results": [
        "Loaded plugins: search-disabled-repos\nResolving Dependencies\n--> Running transaction check\n---> Package libblockdev-crypto.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libblockdev-utils(x86-64) = 2.18-5.el7 for package: libblockdev-crypto-2.18-5.el7.x86_64\n--> Processing Dependency: libvolume_key.so.1()(64bit) for package: libblockdev-crypto-2.18-5.el7.x86_64\n--> Processing Dependency: libbd_utils.so.2()(64bit) for package: libblockdev-crypto-2.18-5.el7.x86_64\n---> Package libblockdev-dm.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libdmraid.so.1(Base)(64bit) for package: libblockdev-dm-2.18-5.el7.x86_64\n--> Processing Dependency: dmraid for package: libblockdev-dm-2.18-5.el7.x86_64\n--> Processing Dependency: libdmraid.so.1()(64bit) for package: libblockdev-dm-2.18-5.el7.x86_64\n---> Package libblockdev-lvm.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: lvm2 for package: libblockdev-lvm-2.18-5.el7.x86_64\n--> Processing Dependency: device-mapper-persistent-data for package: libblockdev-lvm-2.18-5.el7.x86_64\n---> Package libblockdev-mdraid.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: mdadm for package: libblockdev-mdraid-2.18-5.el7.x86_64\n--> Processing Dependency: libbytesize.so.1()(64bit) for package: libblockdev-mdraid-2.18-5.el7.x86_64\n---> Package libblockdev-swap.x86_64 0:2.18-5.el7 will be installed\n---> Package python-enum34.noarch 0:1.0.4-1.el7 will be installed\n---> Package python2-blivet3.noarch 1:3.1.3-3.el7 will be installed\n--> Processing Dependency: blivet3-data = 1:3.1.3-3.el7 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-bytesize >= 0.3 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-blockdev >= 2.17 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: pyparted >= 3.9 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-hawkey for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: lsof for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Running transaction check\n---> Package blivet3-data.noarch 1:3.1.3-3.el7 will be installed\n---> Package device-mapper-persistent-data.x86_64 0:0.8.5-3.el7_9.2 will be installed\n--> Processing Dependency: libaio.so.1(LIBAIO_0.4)(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n--> Processing Dependency: libaio.so.1(LIBAIO_0.1)(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n--> Processing Dependency: libaio.so.1()(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n---> Package dmraid.x86_64 0:1.0.0.rc16-28.el7 will be installed\n--> Processing Dependency: libdevmapper-event.so.1.02(Base)(64bit) for package: dmraid-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: dmraid-events for package: dmraid-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: libdevmapper-event.so.1.02()(64bit) for package: dmraid-1.0.0.rc16-28.el7.x86_64\n---> Package libblockdev-utils.x86_64 0:2.18-5.el7 will be installed\n---> Package libbytesize.x86_64 0:1.2-1.el7 will be installed\n--> Processing Dependency: libmpfr.so.4()(64bit) for package: libbytesize-1.2-1.el7.x86_64\n---> Package lsof.x86_64 0:4.87-6.el7 will be installed\n---> Package lvm2.x86_64 7:2.02.187-6.el7_9.5 will be installed\n--> Processing Dependency: lvm2-libs = 7:2.02.187-6.el7_9.5 for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n--> Processing Dependency: liblvm2app.so.2.2(Base)(64bit) for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n--> Processing Dependency: liblvm2app.so.2.2()(64bit) for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n---> Package mdadm.x86_64 0:4.1-9.el7_9 will be installed\n--> Processing Dependency: libreport-filesystem for package: mdadm-4.1-9.el7_9.x86_64\n---> Package pyparted.x86_64 1:3.9-15.el7 will be installed\n---> Package python2-blockdev.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libblockdev(x86-64) = 2.18-5.el7 for package: python2-blockdev-2.18-5.el7.x86_64\n---> Package python2-bytesize.x86_64 0:1.2-1.el7 will be installed\n---> Package python2-hawkey.x86_64 0:0.22.5-2.el7_9 will be installed\n--> Processing Dependency: libdnf(x86-64) = 0.22.5-2.el7_9 for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: python2-libdnf = 0.22.5-2.el7_9 for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolv.so.0(SOLV_1.0)(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolvext.so.0(SOLV_1.0)(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libdnf.so.2()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libjson-glib-1.0.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libmodulemd.so.1()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: librepo.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: librhsm.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolv.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolvext.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n---> Package volume_key-libs.x86_64 0:0.3.9-9.el7 will be installed\n--> Running transaction check\n---> Package device-mapper-event-libs.x86_64 7:1.02.170-6.el7_9.5 will be installed\n---> Package dmraid-events.x86_64 0:1.0.0.rc16-28.el7 will be installed\n--> Processing Dependency: sgpio for package: dmraid-events-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: device-mapper-event for package: dmraid-events-1.0.0.rc16-28.el7.x86_64\n---> Package json-glib.x86_64 0:1.4.2-2.el7 will be installed\n---> Package libaio.x86_64 0:0.3.109-13.el7 will be installed\n---> Package libblockdev.x86_64 0:2.18-5.el7 will be installed\n---> Package libdnf.x86_64 0:0.22.5-2.el7_9 will be installed\n---> Package libmodulemd.x86_64 0:1.6.3-1.el7 will be installed\n---> Package librepo.x86_64 0:1.8.1-8.el7_9 will be installed\n---> Package libreport-filesystem.x86_64 0:2.1.11-53.el7 will be installed\n---> Package librhsm.x86_64 0:0.0.3-3.el7_9 will be installed\n---> Package libsolv.x86_64 0:0.6.34-4.el7 will be installed\n---> Package lvm2-libs.x86_64 7:2.02.187-6.el7_9.5 will be installed\n---> Package mpfr.x86_64 0:3.1.1-4.el7 will be installed\n---> Package python2-libdnf.x86_64 0:0.22.5-2.el7_9 will be installed\n--> Running transaction check\n---> Package device-mapper-event.x86_64 7:1.02.170-6.el7_9.5 will be installed\n---> Package sgpio.x86_64 0:1.2.0.10-13.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package            Arch   Version              Repository                 Size\n================================================================================\nInstalling:\n libblockdev-crypto x86_64 2.18-5.el7           rhel                       60 k\n libblockdev-dm     x86_64 2.18-5.el7           rhel-optional              54 k\n libblockdev-lvm    x86_64 2.18-5.el7           rhel                       62 k\n libblockdev-mdraid x86_64 2.18-5.el7           rhel                       57 k\n libblockdev-swap   x86_64 2.18-5.el7           rhel                       52 k\n python-enum34      noarch 1.0.4-1.el7          rhel                       52 k\n python2-blivet3    noarch 1:3.1.3-3.el7        rhel                      851 k\nInstalling for dependencies:\n blivet3-data       noarch 1:3.1.3-3.el7        rhel                       77 k\n device-mapper-event\n                    x86_64 7:1.02.170-6.el7_9.5 rhel                      192 k\n device-mapper-event-libs\n                    x86_64 7:1.02.170-6.el7_9.5 rhel                      192 k\n device-mapper-persistent-data\n                    x86_64 0.8.5-3.el7_9.2      rhel                      423 k\n dmraid             x86_64 1.0.0.rc16-28.el7    rhel                      151 k\n dmraid-events      x86_64 1.0.0.rc16-28.el7    rhel                       21 k\n json-glib          x86_64 1.4.2-2.el7          rhel                      134 k\n libaio             x86_64 0.3.109-13.el7       rhel                       24 k\n libblockdev        x86_64 2.18-5.el7           rhel                      119 k\n libblockdev-utils  x86_64 2.18-5.el7           rhel                       59 k\n libbytesize        x86_64 1.2-1.el7            rhel                       52 k\n libdnf             x86_64 0.22.5-2.el7_9       rhel-7-server-extras-rpms 536 k\n libmodulemd        x86_64 1.6.3-1.el7          rhel-7-server-extras-rpms 153 k\n librepo            x86_64 1.8.1-8.el7_9        rhel                       82 k\n libreport-filesystem\n                    x86_64 2.1.11-53.el7        rhel                       41 k\n librhsm            x86_64 0.0.3-3.el7_9        rhel-7-server-extras-rpms  28 k\n libsolv            x86_64 0.6.34-4.el7         rhel                      329 k\n lsof               x86_64 4.87-6.el7           rhel                      331 k\n lvm2               x86_64 7:2.02.187-6.el7_9.5 rhel                      1.3 M\n lvm2-libs          x86_64 7:2.02.187-6.el7_9.5 rhel                      1.1 M\n mdadm              x86_64 4.1-9.el7_9          rhel                      440 k\n mpfr               x86_64 3.1.1-4.el7          rhel                      203 k\n pyparted           x86_64 1:3.9-15.el7         rhel                      195 k\n python2-blockdev   x86_64 2.18-5.el7           rhel                       61 k\n python2-bytesize   x86_64 1.2-1.el7            rhel                       22 k\n python2-hawkey     x86_64 0.22.5-2.el7_9       rhel-7-server-extras-rpms  71 k\n python2-libdnf     x86_64 0.22.5-2.el7_9       rhel-7-server-extras-rpms 611 k\n sgpio              x86_64 1.2.0.10-13.el7      rhel                       14 k\n volume_key-libs    x86_64 0.3.9-9.el7          rhel                      141 k\n\nTransaction Summary\n================================================================================\nInstall  7 Packages (+29 Dependent packages)\n\nTotal download size: 8.2 M\nInstalled size: 24 M\nDownloading packages:\n--------------------------------------------------------------------------------\nTotal                                               16 MB/s | 8.2 MB  00:00     \nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n  Installing : libblockdev-utils-2.18-5.el7.x86_64                         1/36 \n  Installing : 7:device-mapper-event-libs-1.02.170-6.el7_9.5.x86_64        2/36 \n  Installing : json-glib-1.4.2-2.el7.x86_64                                3/36 \n  Installing : librhsm-0.0.3-3.el7_9.x86_64                                4/36 \n  Installing : libsolv-0.6.34-4.el7.x86_64                                 5/36 \n  Installing : libaio-0.3.109-13.el7.x86_64                                6/36 \n  Installing : librepo-1.8.1-8.el7_9.x86_64                                7/36 \n  Installing : libmodulemd-1.6.3-1.el7.x86_64                              8/36 \n  Installing : libdnf-0.22.5-2.el7_9.x86_64                                9/36 \n  Installing : device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64       10/36 \n  Installing : 7:device-mapper-event-1.02.170-6.el7_9.5.x86_64            11/36 \n  Installing : 7:lvm2-libs-2.02.187-6.el7_9.5.x86_64                      12/36 \n  Installing : 7:lvm2-2.02.187-6.el7_9.5.x86_64                           13/36 \n  Installing : python2-libdnf-0.22.5-2.el7_9.x86_64                       14/36 \n  Installing : python2-hawkey-0.22.5-2.el7_9.x86_64                       15/36 \n  Installing : libblockdev-2.18-5.el7.x86_64                              16/36 \n  Installing : python2-blockdev-2.18-5.el7.x86_64                         17/36 \n  Installing : 1:pyparted-3.9-15.el7.x86_64                               18/36 \n  Installing : sgpio-1.2.0.10-13.el7.x86_64                               19/36 \n  Installing : dmraid-1.0.0.rc16-28.el7.x86_64                            20/36 \n  Installing : dmraid-events-1.0.0.rc16-28.el7.x86_64                     21/36 \n  Installing : volume_key-libs-0.3.9-9.el7.x86_64                         22/36 \n  Installing : mpfr-3.1.1-4.el7.x86_64                                    23/36 \n  Installing : libbytesize-1.2-1.el7.x86_64                               24/36 \n  Installing : python2-bytesize-1.2-1.el7.x86_64                          25/36 \n  Installing : libreport-filesystem-2.1.11-53.el7.x86_64                  26/36 \n  Installing : mdadm-4.1-9.el7_9.x86_64                                   27/36 \n  Installing : 1:blivet3-data-3.1.3-3.el7.noarch                          28/36 \n  Installing : lsof-4.87-6.el7.x86_64                                     29/36 \n  Installing : 1:python2-blivet3-3.1.3-3.el7.noarch                       30/36 \n  Installing : libblockdev-mdraid-2.18-5.el7.x86_64                       31/36 \n  Installing : libblockdev-crypto-2.18-5.el7.x86_64                       32/36 \n  Installing : libblockdev-dm-2.18-5.el7.x86_64                           33/36 \n  Installing : libblockdev-lvm-2.18-5.el7.x86_64                          34/36 \n  Installing : libblockdev-swap-2.18-5.el7.x86_64                         35/36 \n  Installing : python-enum34-1.0.4-1.el7.noarch                           36/36 \n  Verifying  : 7:device-mapper-event-1.02.170-6.el7_9.5.x86_64             1/36 \n  Verifying  : libblockdev-swap-2.18-5.el7.x86_64                          2/36 \n  Verifying  : librhsm-0.0.3-3.el7_9.x86_64                                3/36 \n  Verifying  : libblockdev-lvm-2.18-5.el7.x86_64                           4/36 \n  Verifying  : lsof-4.87-6.el7.x86_64                                      5/36 \n  Verifying  : libblockdev-mdraid-2.18-5.el7.x86_64                        6/36 \n  Verifying  : libdnf-0.22.5-2.el7_9.x86_64                                7/36 \n  Verifying  : python-enum34-1.0.4-1.el7.noarch                            8/36 \n  Verifying  : 1:blivet3-data-3.1.3-3.el7.noarch                           9/36 \n  Verifying  : dmraid-events-1.0.0.rc16-28.el7.x86_64                     10/36 \n  Verifying  : python2-blockdev-2.18-5.el7.x86_64                         11/36 \n  Verifying  : libmodulemd-1.6.3-1.el7.x86_64                             12/36 \n  Verifying  : librepo-1.8.1-8.el7_9.x86_64                               13/36 \n  Verifying  : libblockdev-dm-2.18-5.el7.x86_64                           14/36 \n  Verifying  : json-glib-1.4.2-2.el7.x86_64                               15/36 \n  Verifying  : libaio-0.3.109-13.el7.x86_64                               16/36 \n  Verifying  : 7:lvm2-libs-2.02.187-6.el7_9.5.x86_64                      17/36 \n  Verifying  : python2-hawkey-0.22.5-2.el7_9.x86_64                       18/36 \n  Verifying  : python2-bytesize-1.2-1.el7.x86_64                          19/36 \n  Verifying  : libblockdev-2.18-5.el7.x86_64                              20/36 \n  Verifying  : libreport-filesystem-2.1.11-53.el7.x86_64                  21/36 \n  Verifying  : libbytesize-1.2-1.el7.x86_64                               22/36 \n  Verifying  : 7:device-mapper-event-libs-1.02.170-6.el7_9.5.x86_64       23/36 \n  Verifying  : python2-libdnf-0.22.5-2.el7_9.x86_64                       24/36 \n  Verifying  : 7:lvm2-2.02.187-6.el7_9.5.x86_64                           25/36 \n  Verifying  : libblockdev-utils-2.18-5.el7.x86_64                        26/36 \n  Verifying  : mpfr-3.1.1-4.el7.x86_64                                    27/36 \n  Verifying  : volume_key-libs-0.3.9-9.el7.x86_64                         28/36 \n  Verifying  : libsolv-0.6.34-4.el7.x86_64                                29/36 \n  Verifying  : device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64       30/36 \n  Verifying  : 1:python2-blivet3-3.1.3-3.el7.noarch                       31/36 \n  Verifying  : dmraid-1.0.0.rc16-28.el7.x86_64                            32/36 \n  Verifying  : mdadm-4.1-9.el7_9.x86_64                                   33/36 \n  Verifying  : sgpio-1.2.0.10-13.el7.x86_64                               34/36 \n  Verifying  : libblockdev-crypto-2.18-5.el7.x86_64                       35/36 \n  Verifying  : 1:pyparted-3.9-15.el7.x86_64                               36/36 \n\nInstalled:\n  libblockdev-crypto.x86_64 0:2.18-5.el7 libblockdev-dm.x86_64 0:2.18-5.el7    \n  libblockdev-lvm.x86_64 0:2.18-5.el7    libblockdev-mdraid.x86_64 0:2.18-5.el7\n  libblockdev-swap.x86_64 0:2.18-5.el7   python-enum34.noarch 0:1.0.4-1.el7    \n  python2-blivet3.noarch 1:3.1.3-3.el7  \n\nDependency Installed:\n  blivet3-data.noarch 1:3.1.3-3.el7                                             \n  device-mapper-event.x86_64 7:1.02.170-6.el7_9.5                               \n  device-mapper-event-libs.x86_64 7:1.02.170-6.el7_9.5                          \n  device-mapper-persistent-data.x86_64 0:0.8.5-3.el7_9.2                        \n  dmraid.x86_64 0:1.0.0.rc16-28.el7                                             \n  dmraid-events.x86_64 0:1.0.0.rc16-28.el7                                      \n  json-glib.x86_64 0:1.4.2-2.el7                                                \n  libaio.x86_64 0:0.3.109-13.el7                                                \n  libblockdev.x86_64 0:2.18-5.el7                                               \n  libblockdev-utils.x86_64 0:2.18-5.el7                                         \n  libbytesize.x86_64 0:1.2-1.el7                                                \n  libdnf.x86_64 0:0.22.5-2.el7_9                                                \n  libmodulemd.x86_64 0:1.6.3-1.el7                                              \n  librepo.x86_64 0:1.8.1-8.el7_9                                                \n  libreport-filesystem.x86_64 0:2.1.11-53.el7                                   \n  librhsm.x86_64 0:0.0.3-3.el7_9                                                \n  libsolv.x86_64 0:0.6.34-4.el7                                                 \n  lsof.x86_64 0:4.87-6.el7                                                      \n  lvm2.x86_64 7:2.02.187-6.el7_9.5                                              \n  lvm2-libs.x86_64 7:2.02.187-6.el7_9.5                                         \n  mdadm.x86_64 0:4.1-9.el7_9                                                    \n  mpfr.x86_64 0:3.1.1-4.el7                                                     \n  pyparted.x86_64 1:3.9-15.el7                                                  \n  python2-blockdev.x86_64 0:2.18-5.el7                                          \n  python2-bytesize.x86_64 0:1.2-1.el7                                           \n  python2-hawkey.x86_64 0:0.22.5-2.el7_9                                        \n  python2-libdnf.x86_64 0:0.22.5-2.el7_9                                        \n  sgpio.x86_64 0:1.2.0.10-13.el7                                                \n  volume_key-libs.x86_64 0:0.3.9-9.el7                                          \n\nComplete!\n"
    ]
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:39:31 +0000 (0:00:08.905)       0:00:12.018 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:39:31 +0000 (0:00:00.035)       0:00:12.054 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:39:31 +0000 (0:00:00.030)       0:00:12.085 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:39:32 +0000 (0:00:00.646)       0:00:12.731 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : check if the COPR support packages should be installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2
Thursday 21 July 2022  15:39:32 +0000 (0:00:00.042)       0:00:12.774 ********* 

TASK [linux-system-roles.storage : make sure COPR support packages are present] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13
Thursday 21 July 2022  15:39:32 +0000 (0:00:00.031)       0:00:12.805 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable COPRs] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18
Thursday 21 July 2022  15:39:32 +0000 (0:00:00.034)       0:00:12.840 ********* 

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:39:32 +0000 (0:00:00.031)       0:00:12.872 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "rc": 0,
    "results": []
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:39:33 +0000 (0:00:00.531)       0:00:13.403 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "arp-ethers.service": {
                "name": "arp-ethers.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "brandbot.service": {
                "name": "brandbot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "console-shell.service": {
                "name": "console-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "display-manager.service": {
                "name": "display-manager.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dmraid-activation.service": {
                "name": "dmraid-activation.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "exim.service": {
                "name": "exim.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "active"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ip6tables.service": {
                "name": "ip6tables.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "iptables.service": {
                "name": "iptables.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-activation-early.service": {
                "name": "lvm2-activation-early.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "lvm2-activation.service": {
                "name": "lvm2-activation.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "netconsole": {
                "name": "netconsole",
                "source": "sysv",
                "state": "stopped",
                "status": "disabled"
            },
            "network": {
                "name": "network",
                "source": "sysv",
                "state": "running",
                "status": "enabled"
            },
            "network.service": {
                "name": "network.service",
                "source": "systemd",
                "state": "stopped",
                "status": "active"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-config.service": {
                "name": "nfs-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-idmap.service": {
                "name": "nfs-idmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-lock.service": {
                "name": "nfs-lock.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-secure-server.service": {
                "name": "nfs-secure-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "nfs-secure.service": {
                "name": "nfs-secure.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs.service": {
                "name": "nfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfslock.service": {
                "name": "nfslock.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "ntpd.service": {
                "name": "ntpd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "ntpdate.service": {
                "name": "ntpdate.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "ovirt-guest-agent.service": {
                "name": "ovirt-guest-agent.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "postfix.service": {
                "name": "postfix.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rhcd.service": {
                "name": "rhcd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-configure.service": {
                "name": "rhel-configure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-domainname.service": {
                "name": "rhel-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-import-state.service": {
                "name": "rhel-import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-readonly.service": {
                "name": "rhel-readonly.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhnsd": {
                "name": "rhnsd",
                "source": "sysv",
                "state": "running",
                "status": "enabled"
            },
            "rhnsd.service": {
                "name": "rhnsd.service",
                "source": "systemd",
                "state": "running",
                "status": "active"
            },
            "rhsm-facts.service": {
                "name": "rhsm-facts.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rhsm.service": {
                "name": "rhsm.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rhsmcertd.service": {
                "name": "rhsmcertd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpcgssd.service": {
                "name": "rpcgssd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rpcidmapd.service": {
                "name": "rpcidmapd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rsyncd.service": {
                "name": "rsyncd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rsyncd@.service": {
                "name": "rsyncd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service",
                "source": "systemd",
                "state": "stopped",
                "status": "active"
            },
            "sendmail.service": {
                "name": "sendmail.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service",
                "source": "systemd",
                "state": "running",
                "status": "active"
            },
            "sntp.service": {
                "name": "sntp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "sshd-keygen.service": {
                "name": "sshd-keygen.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-importd.service": {
                "name": "systemd-importd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-machined.service": {
                "name": "systemd-machined.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-timesyncd.service": {
                "name": "systemd-timesyncd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ypbind.service": {
                "name": "ypbind.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "yppasswdd.service": {
                "name": "yppasswdd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "ypserv.service": {
                "name": "ypserv.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "ypxfrd.service": {
                "name": "ypxfrd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            }
        }
    },
    "changed": false
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:39:34 +0000 (0:00:01.134)       0:00:14.537 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:39:34 +0000 (0:00:00.111)       0:00:14.649 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:39:34 +0000 (0:00:00.020)       0:00:14.669 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:39:34 +0000 (0:00:00.468)       0:00:15.138 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:39:34 +0000 (0:00:00.035)       0:00:15.173 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:39:34 +0000 (0:00:00.021)       0:00:15.194 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [],
        "changed": false,
        "crypts": [],
        "failed": false,
        "leaves": [],
        "mounts": [],
        "packages": [],
        "pools": [],
        "volumes": []
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:39:35 +0000 (0:00:00.034)       0:00:15.229 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:39:35 +0000 (0:00:00.036)       0:00:15.266 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:39:35 +0000 (0:00:00.036)       0:00:15.302 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:39:35 +0000 (0:00:00.034)       0:00:15.337 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:39:35 +0000 (0:00:00.023)       0:00:15.360 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:39:35 +0000 (0:00:00.036)       0:00:15.396 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:39:35 +0000 (0:00:00.023)       0:00:15.420 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658417969.482065,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:39:35 +0000 (0:00:00.443)       0:00:15.864 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:39:35 +0000 (0:00:00.022)       0:00:15.886 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [Mark tasks to be skipped] ************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:19
Thursday 21 July 2022  15:39:36 +0000 (0:00:00.826)       0:00:16.712 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_skip_checks": [
            "blivet_available",
            "packages_installed",
            "service_facts"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:26
Thursday 21 July 2022  15:39:36 +0000 (0:00:00.033)       0:00:16.746 ********* 
included: /tmp/tmpaxjje44y/tests/get_unused_disk.yml for /cache/rhel-7.qcow2

TASK [Find unused disks in the system] *****************************************
task path: /tmp/tmpaxjje44y/tests/get_unused_disk.yml:2
Thursday 21 July 2022  15:39:36 +0000 (0:00:00.033)       0:00:16.780 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "disks": [
        "nvme1n1"
    ]
}

TASK [Set unused_disks if necessary] *******************************************
task path: /tmp/tmpaxjje44y/tests/get_unused_disk.yml:9
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.455)       0:00:17.235 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "unused_disks": [
            "nvme1n1"
        ]
    },
    "changed": false
}

TASK [Exit playbook when there's not enough unused disks in the system] ********
task path: /tmp/tmpaxjje44y/tests/get_unused_disk.yml:14
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.066)       0:00:17.301 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Print unused disks] ******************************************************
task path: /tmp/tmpaxjje44y/tests/get_unused_disk.yml:19
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.038)       0:00:17.340 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "unused_disks": [
        "nvme1n1"
    ]
}

TASK [Test creating ext4 filesystem with valid parameter "-Fb 4096"] ***********
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:31
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.032)       0:00:17.372 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.040)       0:00:17.413 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.058)       0:00:17.471 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.407)       0:00:17.878 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.056)       0:00:17.935 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.029)       0:00:17.965 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.030)       0:00:17.995 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.042)       0:00:18.038 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.020)       0:00:18.058 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.028)       0:00:18.086 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_create_options": "-Fb 4096",
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.038)       0:00:18.125 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.034)       0:00:18.160 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.027)       0:00:18.188 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:39:37 +0000 (0:00:00.027)       0:00:18.215 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:39:38 +0000 (0:00:00.031)       0:00:18.246 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:39:38 +0000 (0:00:00.025)       0:00:18.272 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:39:38 +0000 (0:00:00.058)       0:00:18.330 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:39:38 +0000 (0:00:00.020)       0:00:18.351 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/nvme1n1",
            "fs_type": "lvmpv"
        },
        {
            "action": "create device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "create device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd",
        "/dev/mapper/foo-test1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "opts": "defaults",
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs",
        "lvm2"
    ],
    "pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "-Fb 4096",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:39:39 +0000 (0:00:01.657)       0:00:20.008 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:39:39 +0000 (0:00:00.037)       0:00:20.045 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:39:39 +0000 (0:00:00.022)       0:00:20.068 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/nvme1n1",
                "fs_type": "lvmpv"
            },
            {
                "action": "create device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "create device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd",
            "/dev/mapper/foo-test1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "opts": "defaults",
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs",
            "lvm2"
        ],
        "pools": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "-Fb 4096",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:39:39 +0000 (0:00:00.038)       0:00:20.106 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "-Fb 4096",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:39:39 +0000 (0:00:00.038)       0:00:20.145 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:39:40 +0000 (0:00:00.075)       0:00:20.220 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:39:40 +0000 (0:00:00.043)       0:00:20.264 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:39:40 +0000 (0:00:00.714)       0:00:20.978 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'ext4', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "opts": "defaults",
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:39:41 +0000 (0:00:00.472)       0:00:21.451 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:39:41 +0000 (0:00:00.474)       0:00:21.926 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658417969.482065,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:39:42 +0000 (0:00:00.326)       0:00:22.252 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:39:42 +0000 (0:00:00.022)       0:00:22.275 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:46
Thursday 21 July 2022  15:39:43 +0000 (0:00:01.850)       0:00:24.125 ********* 
included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1
Thursday 21 July 2022  15:39:43 +0000 (0:00:00.041)       0:00:24.167 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "-Fb 4096",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6
Thursday 21 July 2022  15:39:44 +0000 (0:00:00.051)       0:00:24.219 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14
Thursday 21 July 2022  15:39:44 +0000 (0:00:00.036)       0:00:24.255 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/mapper/foo-test1": {
            "fstype": "ext4",
            "label": "",
            "name": "/dev/mapper/foo-test1",
            "size": "4G",
            "type": "lvm",
            "uuid": "24adbeed-e4e5-4398-8e6e-7d782884bda1"
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "LVM2_member",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": "ypslHj-OxCa-4rB8-mPy1-ijhY-9wRB-g0BXYo"
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-15-39-07-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19
Thursday 21 July 2022  15:39:44 +0000 (0:00:00.464)       0:00:24.720 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003392",
    "end": "2022-07-21 11:39:44.929318",
    "rc": 0,
    "start": "2022-07-21 11:39:44.925926"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0
/dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24
Thursday 21 July 2022  15:39:44 +0000 (0:00:00.450)       0:00:25.170 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003358",
    "end": "2022-07-21 11:39:45.232685",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 11:39:45.229327"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33
Thursday 21 July 2022  15:39:45 +0000 (0:00:00.307)       0:00:25.477 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None})

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:5
Thursday 21 July 2022  15:39:45 +0000 (0:00:00.057)       0:00:25.535 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:18
Thursday 21 July 2022  15:39:45 +0000 (0:00:00.036)       0:00:25.571 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members)
included: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes)

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:1
Thursday 21 July 2022  15:39:45 +0000 (0:00:00.046)       0:00:25.618 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/nvme1n1"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:10
Thursday 21 July 2022  15:39:45 +0000 (0:00:00.051)       0:00:25.670 ********* 
ok: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/nvme1n1",
    "pv": "/dev/nvme1n1"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:19
Thursday 21 July 2022  15:39:45 +0000 (0:00:00.430)       0:00:26.100 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:23
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.119)       0:00:26.220 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/nvme1n1"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:27
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.060)       0:00:26.281 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:34
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.053)       0:00:26.335 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:38
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.039)       0:00:26.374 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:42
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.048)       0:00:26.423 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:46
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.026)       0:00:26.449 ********* 
ok: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/nvme1n1"
}

MSG:

All assertions passed

TASK [Check MD RAID] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:56
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.044)       0:00:26.494 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-md.yml for /cache/rhel-7.qcow2

TASK [get information about RAID] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:6
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.044)       0:00:26.538 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:12
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.023)       0:00:26.561 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:16
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.026)       0:00:26.588 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:20
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.023)       0:00:26.611 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:24
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.025)       0:00:26.636 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:30
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.025)       0:00:26.662 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:36
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.026)       0:00:26.689 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:44
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.023)       0:00:26.712 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:59
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.032)       0:00:26.745 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml:1
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.044)       0:00:26.789 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'})

TASK [Get information about LVM RAID] ******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.041)       0:00:26.831 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is LVM RAID] *******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.040)       0:00:26.872 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.041)       0:00:26.913 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:62
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.030)       0:00:26.943 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml for /cache/rhel-7.qcow2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml:1
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.045)       0:00:26.989 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'})

TASK [Get information about thinpool] ******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.045)       0:00:27.035 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.024)       0:00:27.060 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.022)       0:00:27.082 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.023)       0:00:27.106 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check member encryption] *************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:65
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.024)       0:00:27.130 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:4
Thursday 21 July 2022  15:39:46 +0000 (0:00:00.046)       0:00:27.177 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:8
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.048)       0:00:27.225 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1)  => {
    "_storage_test_pool_member_path": "/dev/nvme1n1",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:15
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.025)       0:00:27.250 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/nvme1n1)

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:1
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.043)       0:00:27.294 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:6
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.047)       0:00:27.341 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:11
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.079)       0:00:27.421 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:17
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.034)       0:00:27.456 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:23
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.101)       0:00:27.557 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:29
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.035)       0:00:27.593 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:22
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.034)       0:00:27.627 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:68
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.035)       0:00:27.663 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml:1
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.047)       0:00:27.710 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'})

TASK [get information about VDO deduplication] *********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.045)       0:00:27.756 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.022)       0:00:27.778 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.022)       0:00:27.801 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.022)       0:00:27.823 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.023)       0:00:27.846 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.022)       0:00:27.868 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.022)       0:00:27.891 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.026)       0:00:27.918 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:71
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.046)       0:00:27.965 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [verify the volumes] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml:3
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.039)       0:00:28.004 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'})

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.042)       0:00:28.046 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.048)       0:00:28.095 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache)

TASK [Get expected mount device based on device type] **************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.077)       0:00:28.173 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14
Thursday 21 July 2022  15:39:47 +0000 (0:00:00.039)       0:00:28.213 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": [
            {
                "block_available": 938726,
                "block_size": 4096,
                "block_total": 999344,
                "block_used": 60618,
                "device": "/dev/mapper/foo-test1",
                "fstype": "ext4",
                "inode_available": 262133,
                "inode_total": 262144,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 3845021696,
                "size_total": 4093313024,
                "uuid": "24adbeed-e4e5-4398-8e6e-7d782884bda1"
            }
        ],
        "storage_test_mount_expected_match_count": "1",
        "storage_test_mount_point_matches": [
            {
                "block_available": 938726,
                "block_size": 4096,
                "block_total": 999344,
                "block_used": 60618,
                "device": "/dev/mapper/foo-test1",
                "fstype": "ext4",
                "inode_available": 262133,
                "inode_total": 262144,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 3845021696,
                "size_total": 4093313024,
                "uuid": "24adbeed-e4e5-4398-8e6e-7d782884bda1"
            }
        ],
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28
Thursday 21 July 2022  15:39:48 +0000 (0:00:00.058)       0:00:28.271 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the current mount state by mount point] ***************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37
Thursday 21 July 2022  15:39:48 +0000 (0:00:00.058)       0:00:28.330 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the mount fs type] ************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45
Thursday 21 July 2022  15:39:48 +0000 (0:00:00.048)       0:00:28.379 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [command] *****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54
Thursday 21 July 2022  15:39:48 +0000 (0:00:00.049)       0:00:28.428 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58
Thursday 21 July 2022  15:39:48 +0000 (0:00:00.022)       0:00:28.451 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63
Thursday 21 July 2022  15:39:48 +0000 (0:00:00.022)       0:00:28.474 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75
Thursday 21 July 2022  15:39:48 +0000 (0:00:00.022)       0:00:28.496 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": null,
        "storage_test_mount_expected_match_count": null,
        "storage_test_mount_point_matches": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2
Thursday 21 July 2022  15:39:48 +0000 (0:00:00.034)       0:00:28.531 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25
Thursday 21 July 2022  15:39:48 +0000 (0:00:00.059)       0:00:28.590 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32
Thursday 21 July 2022  15:39:48 +0000 (0:00:00.048)       0:00:28.639 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39
Thursday 21 July 2022  15:39:48 +0000 (0:00:00.052)       0:00:28.691 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up variables] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49
Thursday 21 July 2022  15:39:48 +0000 (0:00:00.035)       0:00:28.727 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4
Thursday 21 July 2022  15:39:48 +0000 (0:00:00.065)       0:00:28.792 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10
Thursday 21 July 2022  15:39:48 +0000 (0:00:00.118)       0:00:28.911 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4
Thursday 21 July 2022  15:39:48 +0000 (0:00:00.038)       0:00:28.950 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658417979.724065,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1658417979.724065,
        "dev": 5,
        "device_type": 64512,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 34835,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1658417979.724065,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10
Thursday 21 July 2022  15:39:49 +0000 (0:00:00.332)       0:00:29.282 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18
Thursday 21 July 2022  15:39:49 +0000 (0:00:00.037)       0:00:29.319 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [(1/2) Process volume type (set initial value)] ***************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24
Thursday 21 July 2022  15:39:49 +0000 (0:00:00.039)       0:00:29.358 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [(2/2) Process volume type (get RAID value)] ******************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28
Thursday 21 July 2022  15:39:49 +0000 (0:00:00.035)       0:00:29.394 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33
Thursday 21 July 2022  15:39:49 +0000 (0:00:00.024)       0:00:29.418 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3
Thursday 21 July 2022  15:39:49 +0000 (0:00:00.038)       0:00:29.457 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10
Thursday 21 July 2022  15:39:49 +0000 (0:00:00.024)       0:00:29.481 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "changed": true,
    "changes": {
        "installed": [
            "cryptsetup"
        ]
    },
    "rc": 0,
    "results": [
        "Loaded plugins: search-disabled-repos\nResolving Dependencies\n--> Running transaction check\n---> Package cryptsetup.x86_64 0:2.0.3-6.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package             Arch            Version                Repository     Size\n================================================================================\nInstalling:\n cryptsetup          x86_64          2.0.3-6.el7            rhel          154 k\n\nTransaction Summary\n================================================================================\nInstall  1 Package\n\nTotal download size: 154 k\nInstalled size: 354 k\nDownloading packages:\nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n  Installing : cryptsetup-2.0.3-6.el7.x86_64                                1/1 \n  Verifying  : cryptsetup-2.0.3-6.el7.x86_64                                1/1 \n\nInstalled:\n  cryptsetup.x86_64 0:2.0.3-6.el7                                               \n\nComplete!\n"
    ]
}

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15
Thursday 21 July 2022  15:39:50 +0000 (0:00:01.510)       0:00:30.992 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21
Thursday 21 July 2022  15:39:50 +0000 (0:00:00.023)       0:00:31.015 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30
Thursday 21 July 2022  15:39:50 +0000 (0:00:00.024)       0:00:31.040 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38
Thursday 21 July 2022  15:39:50 +0000 (0:00:00.051)       0:00:31.092 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44
Thursday 21 July 2022  15:39:50 +0000 (0:00:00.025)       0:00:31.117 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49
Thursday 21 July 2022  15:39:50 +0000 (0:00:00.024)       0:00:31.141 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55
Thursday 21 July 2022  15:39:50 +0000 (0:00:00.025)       0:00:31.167 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61
Thursday 21 July 2022  15:39:50 +0000 (0:00:00.025)       0:00:31.192 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67
Thursday 21 July 2022  15:39:50 +0000 (0:00:00.026)       0:00:31.218 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74
Thursday 21 July 2022  15:39:51 +0000 (0:00:00.050)       0:00:31.269 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79
Thursday 21 July 2022  15:39:51 +0000 (0:00:00.050)       0:00:31.319 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85
Thursday 21 July 2022  15:39:51 +0000 (0:00:00.039)       0:00:31.359 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91
Thursday 21 July 2022  15:39:51 +0000 (0:00:00.042)       0:00:31.402 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97
Thursday 21 July 2022  15:39:51 +0000 (0:00:00.045)       0:00:31.448 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [get information about RAID] **********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7
Thursday 21 July 2022  15:39:51 +0000 (0:00:00.037)       0:00:31.486 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13
Thursday 21 July 2022  15:39:51 +0000 (0:00:00.043)       0:00:31.529 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17
Thursday 21 July 2022  15:39:51 +0000 (0:00:00.037)       0:00:31.567 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21
Thursday 21 July 2022  15:39:51 +0000 (0:00:00.037)       0:00:31.605 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25
Thursday 21 July 2022  15:39:51 +0000 (0:00:00.037)       0:00:31.643 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31
Thursday 21 July 2022  15:39:51 +0000 (0:00:00.041)       0:00:31.685 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37
Thursday 21 July 2022  15:39:51 +0000 (0:00:00.037)       0:00:31.722 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the actual size of the volume] *************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3
Thursday 21 July 2022  15:39:51 +0000 (0:00:00.035)       0:00:31.758 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [parse the requested size of the volume] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9
Thursday 21 July 2022  15:39:51 +0000 (0:00:00.437)       0:00:32.195 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15
Thursday 21 July 2022  15:39:52 +0000 (0:00:00.365)       0:00:32.561 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_expected_size": "4294967296"
    },
    "changed": false
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20
Thursday 21 July 2022  15:39:52 +0000 (0:00:00.111)       0:00:32.672 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25
Thursday 21 July 2022  15:39:52 +0000 (0:00:00.076)       0:00:32.749 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28
Thursday 21 July 2022  15:39:52 +0000 (0:00:00.037)       0:00:32.787 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31
Thursday 21 July 2022  15:39:52 +0000 (0:00:00.035)       0:00:32.822 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36
Thursday 21 July 2022  15:39:52 +0000 (0:00:00.039)       0:00:32.862 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39
Thursday 21 July 2022  15:39:52 +0000 (0:00:00.036)       0:00:32.898 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44
Thursday 21 July 2022  15:39:52 +0000 (0:00:00.038)       0:00:32.936 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_actual_size": {
        "bytes": 4294967296,
        "changed": false,
        "failed": false,
        "lvm": "4g",
        "parted": "4GiB",
        "size": "4 GiB"
    }
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47
Thursday 21 July 2022  15:39:52 +0000 (0:00:00.037)       0:00:32.973 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50
Thursday 21 July 2022  15:39:52 +0000 (0:00:00.039)       0:00:33.013 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6
Thursday 21 July 2022  15:39:52 +0000 (0:00:00.052)       0:00:33.066 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.027107",
    "end": "2022-07-21 11:39:53.187322",
    "rc": 0,
    "start": "2022-07-21 11:39:53.160215"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14
Thursday 21 July 2022  15:39:53 +0000 (0:00:00.365)       0:00:33.432 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [check segment type] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17
Thursday 21 July 2022  15:39:53 +0000 (0:00:00.052)       0:00:33.484 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22
Thursday 21 July 2022  15:39:53 +0000 (0:00:00.051)       0:00:33.536 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested cache size] ******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26
Thursday 21 July 2022  15:39:53 +0000 (0:00:00.036)       0:00:33.573 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32
Thursday 21 July 2022  15:39:53 +0000 (0:00:00.034)       0:00:33.608 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36
Thursday 21 July 2022  15:39:53 +0000 (0:00:00.037)       0:00:33.646 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16
Thursday 21 July 2022  15:39:53 +0000 (0:00:00.035)       0:00:33.682 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43
Thursday 21 July 2022  15:39:53 +0000 (0:00:00.033)       0:00:33.715 ********* 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53
Thursday 21 July 2022  15:39:53 +0000 (0:00:00.020)       0:00:33.735 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Remove the volume group created above] ***********************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:48
Thursday 21 July 2022  15:39:53 +0000 (0:00:00.034)       0:00:33.769 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:39:53 +0000 (0:00:00.048)       0:00:33.818 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:39:53 +0000 (0:00:00.036)       0:00:33.854 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:39:54 +0000 (0:00:00.413)       0:00:34.268 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:39:54 +0000 (0:00:00.062)       0:00:34.330 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:39:54 +0000 (0:00:00.035)       0:00:34.366 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:39:54 +0000 (0:00:00.036)       0:00:34.402 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:39:54 +0000 (0:00:00.047)       0:00:34.450 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:39:54 +0000 (0:00:00.020)       0:00:34.471 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:39:54 +0000 (0:00:00.026)       0:00:34.497 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "state": "absent",
            "type": "lvm"
        }
    ]
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:39:54 +0000 (0:00:00.101)       0:00:34.598 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:39:54 +0000 (0:00:00.036)       0:00:34.635 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:39:54 +0000 (0:00:00.029)       0:00:34.664 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:39:54 +0000 (0:00:00.029)       0:00:34.694 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:39:54 +0000 (0:00:00.028)       0:00:34.722 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:39:54 +0000 (0:00:00.029)       0:00:34.751 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:39:54 +0000 (0:00:00.062)       0:00:34.814 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:39:54 +0000 (0:00:00.025)       0:00:34.840 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/nvme1n1",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "absent",
            "type": "lvm",
            "volumes": []
        }
    ],
    "volumes": []
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:39:56 +0000 (0:00:01.602)       0:00:36.443 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:39:56 +0000 (0:00:00.035)       0:00:36.478 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:39:56 +0000 (0:00:00.020)       0:00:36.499 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/nvme1n1",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ],
        "volumes": []
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:39:56 +0000 (0:00:00.040)       0:00:36.539 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ]
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:39:56 +0000 (0:00:00.037)       0:00:36.576 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:39:56 +0000 (0:00:00.036)       0:00:36.613 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:39:56 +0000 (0:00:00.348)       0:00:36.962 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:39:57 +0000 (0:00:00.464)       0:00:37.427 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:39:57 +0000 (0:00:00.038)       0:00:37.465 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:39:57 +0000 (0:00:00.444)       0:00:37.910 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658417969.482065,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:39:58 +0000 (0:00:00.313)       0:00:38.224 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:39:58 +0000 (0:00:00.024)       0:00:38.248 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:58
Thursday 21 July 2022  15:39:58 +0000 (0:00:00.848)       0:00:39.097 ********* 
included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1
Thursday 21 July 2022  15:39:58 +0000 (0:00:00.040)       0:00:39.138 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "absent",
            "type": "lvm",
            "volumes": []
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6
Thursday 21 July 2022  15:39:59 +0000 (0:00:00.094)       0:00:39.232 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14
Thursday 21 July 2022  15:39:59 +0000 (0:00:00.070)       0:00:39.303 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-15-39-07-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19
Thursday 21 July 2022  15:39:59 +0000 (0:00:00.316)       0:00:39.619 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.004092",
    "end": "2022-07-21 11:39:59.695514",
    "rc": 0,
    "start": "2022-07-21 11:39:59.691422"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24
Thursday 21 July 2022  15:39:59 +0000 (0:00:00.321)       0:00:39.941 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003458",
    "end": "2022-07-21 11:40:00.008055",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 11:40:00.004597"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.310)       0:00:40.252 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'absent', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None})

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:5
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.058)       0:00:40.310 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:18
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.032)       0:00:40.343 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members)
included: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes)

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:1
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.043)       0:00:40.387 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "0",
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:10
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.050)       0:00:40.438 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:19
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.021)       0:00:40.459 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": "0"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:23
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.051)       0:00:40.511 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": []
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:27
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.049)       0:00:40.560 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:34
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.065)       0:00:40.626 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:38
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.035)       0:00:40.661 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:42
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.047)       0:00:40.709 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:46
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.024)       0:00:40.733 ********* 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:56
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.021)       0:00:40.754 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-md.yml for /cache/rhel-7.qcow2

TASK [get information about RAID] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:6
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.041)       0:00:40.796 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:12
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.025)       0:00:40.821 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:16
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.022)       0:00:40.844 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:20
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.026)       0:00:40.870 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:24
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.025)       0:00:40.896 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:30
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.025)       0:00:40.921 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:36
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.022)       0:00:40.944 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:44
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.023)       0:00:40.967 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:59
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.032)       0:00:40.999 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml:1
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.048)       0:00:41.048 ********* 

TASK [Check Thin Pools] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:62
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.023)       0:00:41.072 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml for /cache/rhel-7.qcow2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml:1
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.077)       0:00:41.149 ********* 

TASK [Check member encryption] *************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:65
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.021)       0:00:41.171 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:4
Thursday 21 July 2022  15:40:00 +0000 (0:00:00.045)       0:00:41.216 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:8
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.050)       0:00:41.266 ********* 

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:15
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.024)       0:00:41.291 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:22
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.022)       0:00:41.313 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:68
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.035)       0:00:41.349 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml:1
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.048)       0:00:41.397 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:71
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.023)       0:00:41.420 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [verify the volumes] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml:3
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.036)       0:00:41.457 ********* 

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.023)       0:00:41.480 ********* 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.023)       0:00:41.503 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Try to create ext4 filesystem with invalid parameter "-Fb 512"] **********
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:62
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.031)       0:00:41.535 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.035)       0:00:41.571 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.034)       0:00:41.606 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.431)       0:00:42.038 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.063)       0:00:42.101 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.032)       0:00:42.133 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.035)       0:00:42.168 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:40:01 +0000 (0:00:00.047)       0:00:42.216 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:40:02 +0000 (0:00:00.020)       0:00:42.237 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:40:02 +0000 (0:00:00.029)       0:00:42.266 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_create_options": "-Fb 512",
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:40:02 +0000 (0:00:00.040)       0:00:42.307 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:40:02 +0000 (0:00:00.036)       0:00:42.343 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:40:02 +0000 (0:00:00.026)       0:00:42.370 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:40:02 +0000 (0:00:00.025)       0:00:42.395 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:40:02 +0000 (0:00:00.026)       0:00:42.421 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:40:02 +0000 (0:00:00.025)       0:00:42.447 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:40:02 +0000 (0:00:00.095)       0:00:42.543 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:40:02 +0000 (0:00:00.021)       0:00:42.565 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "xfsprogs",
        "e2fsprogs",
        "lvm2"
    ],
    "pools": [],
    "volumes": []
}

MSG:

Failed to commit changes to disk: (FSError('format failed: 1',), '/dev/mapper/foo-test1')

TASK [linux-system-roles.storage : failed message] *****************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99
Thursday 21 July 2022  15:40:03 +0000 (0:00:01.428)       0:00:43.993 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "changed": false
}

MSG:

{'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': False, 'pools': [{'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 512'}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}], 'volumes': [], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': ['xfsprogs', 'e2fsprogs', 'lvm2'], 'msg': "Failed to commit changes to disk: (FSError('format failed: 1',), '/dev/mapper/foo-test1')", '_ansible_no_log': False}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:40:03 +0000 (0:00:00.039)       0:00:44.032 ********* 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:82
Thursday 21 July 2022  15:40:03 +0000 (0:00:00.020)       0:00:44.053 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the output when creating ext4 filesystem with invalid parameter "-Fb 512"] ***
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:88
Thursday 21 July 2022  15:40:03 +0000 (0:00:00.044)       0:00:44.098 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Remove the volume group created above] ***********************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:95
Thursday 21 July 2022  15:40:03 +0000 (0:00:00.047)       0:00:44.145 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:40:03 +0000 (0:00:00.032)       0:00:44.178 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:40:03 +0000 (0:00:00.031)       0:00:44.210 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:40:04 +0000 (0:00:00.416)       0:00:44.626 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:40:04 +0000 (0:00:00.059)       0:00:44.686 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:40:04 +0000 (0:00:00.035)       0:00:44.721 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:40:04 +0000 (0:00:00.041)       0:00:44.763 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:40:04 +0000 (0:00:00.047)       0:00:44.810 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:40:04 +0000 (0:00:00.019)       0:00:44.829 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:40:04 +0000 (0:00:00.029)       0:00:44.859 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "state": "absent",
            "type": "lvm"
        }
    ]
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:40:04 +0000 (0:00:00.034)       0:00:44.893 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:40:04 +0000 (0:00:00.038)       0:00:44.932 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:40:04 +0000 (0:00:00.032)       0:00:44.965 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:40:04 +0000 (0:00:00.036)       0:00:45.001 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:40:04 +0000 (0:00:00.030)       0:00:45.031 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:40:04 +0000 (0:00:00.030)       0:00:45.062 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:40:04 +0000 (0:00:00.061)       0:00:45.123 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:40:04 +0000 (0:00:00.021)       0:00:45.145 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/nvme1n1",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "absent",
            "type": "lvm",
            "volumes": []
        }
    ],
    "volumes": []
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:40:06 +0000 (0:00:01.455)       0:00:46.600 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:40:06 +0000 (0:00:00.038)       0:00:46.639 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:40:06 +0000 (0:00:00.023)       0:00:46.663 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/nvme1n1",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ],
        "volumes": []
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:40:06 +0000 (0:00:00.081)       0:00:46.745 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ]
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:40:06 +0000 (0:00:00.085)       0:00:46.830 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:40:06 +0000 (0:00:00.038)       0:00:46.869 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:40:06 +0000 (0:00:00.039)       0:00:46.908 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:40:06 +0000 (0:00:00.026)       0:00:46.935 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:40:06 +0000 (0:00:00.035)       0:00:46.970 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:40:06 +0000 (0:00:00.026)       0:00:46.997 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658417969.482065,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:40:07 +0000 (0:00:00.326)       0:00:47.324 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:40:07 +0000 (0:00:00.023)       0:00:47.347 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [Create one LVM logical volume with "4g" under one volume group] **********
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:105
Thursday 21 July 2022  15:40:07 +0000 (0:00:00.824)       0:00:48.172 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:40:07 +0000 (0:00:00.040)       0:00:48.213 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:40:08 +0000 (0:00:00.032)       0:00:48.246 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:40:08 +0000 (0:00:00.429)       0:00:48.675 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:40:08 +0000 (0:00:00.064)       0:00:48.739 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:40:08 +0000 (0:00:00.035)       0:00:48.775 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:40:08 +0000 (0:00:00.035)       0:00:48.810 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:40:08 +0000 (0:00:00.044)       0:00:48.855 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:40:08 +0000 (0:00:00.020)       0:00:48.875 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:40:08 +0000 (0:00:00.026)       0:00:48.902 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:40:08 +0000 (0:00:00.070)       0:00:48.972 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:40:08 +0000 (0:00:00.033)       0:00:49.006 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:40:08 +0000 (0:00:00.029)       0:00:49.036 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:40:08 +0000 (0:00:00.029)       0:00:49.065 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:40:08 +0000 (0:00:00.072)       0:00:49.138 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:40:08 +0000 (0:00:00.029)       0:00:49.168 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:40:09 +0000 (0:00:00.057)       0:00:49.225 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:40:09 +0000 (0:00:00.025)       0:00:49.250 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/nvme1n1",
            "fs_type": "lvmpv"
        },
        {
            "action": "create device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "create device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd",
        "/dev/mapper/foo-test1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "opts": "defaults",
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs",
        "lvm2"
    ],
    "pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:40:10 +0000 (0:00:01.610)       0:00:50.861 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:40:10 +0000 (0:00:00.039)       0:00:50.900 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:40:10 +0000 (0:00:00.021)       0:00:50.922 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/nvme1n1",
                "fs_type": "lvmpv"
            },
            {
                "action": "create device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "create device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd",
            "/dev/mapper/foo-test1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "opts": "defaults",
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs",
            "lvm2"
        ],
        "pools": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:40:10 +0000 (0:00:00.041)       0:00:50.964 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:40:10 +0000 (0:00:00.040)       0:00:51.004 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:40:10 +0000 (0:00:00.040)       0:00:51.044 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:40:10 +0000 (0:00:00.040)       0:00:51.085 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:40:11 +0000 (0:00:00.475)       0:00:51.560 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'ext4', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "opts": "defaults",
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:40:11 +0000 (0:00:00.364)       0:00:51.925 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:40:12 +0000 (0:00:00.483)       0:00:52.408 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658417969.482065,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:40:12 +0000 (0:00:00.308)       0:00:52.717 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:40:12 +0000 (0:00:00.023)       0:00:52.740 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:119
Thursday 21 July 2022  15:40:13 +0000 (0:00:00.825)       0:00:53.566 ********* 
included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1
Thursday 21 July 2022  15:40:13 +0000 (0:00:00.038)       0:00:53.605 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6
Thursday 21 July 2022  15:40:13 +0000 (0:00:00.054)       0:00:53.660 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14
Thursday 21 July 2022  15:40:13 +0000 (0:00:00.041)       0:00:53.702 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/mapper/foo-test1": {
            "fstype": "ext4",
            "label": "",
            "name": "/dev/mapper/foo-test1",
            "size": "4G",
            "type": "lvm",
            "uuid": "f7dcf85c-126f-4bdd-83f1-93b5c5721fc4"
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "LVM2_member",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": "2pk9UR-qCcN-aY1m-zdWe-dtHD-S8NW-O2tcrO"
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-15-39-07-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19
Thursday 21 July 2022  15:40:13 +0000 (0:00:00.328)       0:00:54.030 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.004274",
    "end": "2022-07-21 11:40:14.176868",
    "rc": 0,
    "start": "2022-07-21 11:40:14.172594"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0
/dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24
Thursday 21 July 2022  15:40:14 +0000 (0:00:00.399)       0:00:54.429 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003366",
    "end": "2022-07-21 11:40:14.562500",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 11:40:14.559134"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33
Thursday 21 July 2022  15:40:14 +0000 (0:00:00.376)       0:00:54.806 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None})

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:5
Thursday 21 July 2022  15:40:14 +0000 (0:00:00.061)       0:00:54.867 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:18
Thursday 21 July 2022  15:40:14 +0000 (0:00:00.108)       0:00:54.976 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members)
included: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes)

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:1
Thursday 21 July 2022  15:40:14 +0000 (0:00:00.048)       0:00:55.025 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/nvme1n1"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:10
Thursday 21 July 2022  15:40:14 +0000 (0:00:00.059)       0:00:55.084 ********* 
ok: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/nvme1n1",
    "pv": "/dev/nvme1n1"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:19
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.311)       0:00:55.395 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:23
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.048)       0:00:55.444 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/nvme1n1"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:27
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.053)       0:00:55.497 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:34
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.051)       0:00:55.549 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:38
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.037)       0:00:55.586 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:42
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.049)       0:00:55.636 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:46
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.025)       0:00:55.661 ********* 
ok: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/nvme1n1"
}

MSG:

All assertions passed

TASK [Check MD RAID] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:56
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.041)       0:00:55.702 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-md.yml for /cache/rhel-7.qcow2

TASK [get information about RAID] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:6
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.042)       0:00:55.745 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:12
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.027)       0:00:55.772 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:16
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.024)       0:00:55.796 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:20
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.022)       0:00:55.818 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:24
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.022)       0:00:55.841 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:30
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.023)       0:00:55.865 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:36
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.022)       0:00:55.888 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:44
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.022)       0:00:55.910 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:59
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.041)       0:00:55.952 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml:1
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.047)       0:00:55.999 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [Get information about LVM RAID] ******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.041)       0:00:56.041 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is LVM RAID] *******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.027)       0:00:56.069 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.028)       0:00:56.098 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:62
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.031)       0:00:56.129 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml for /cache/rhel-7.qcow2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml:1
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.043)       0:00:56.173 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [Get information about thinpool] ******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3
Thursday 21 July 2022  15:40:15 +0000 (0:00:00.042)       0:00:56.216 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.022)       0:00:56.238 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.023)       0:00:56.262 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.022)       0:00:56.284 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check member encryption] *************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:65
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.026)       0:00:56.310 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:4
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.048)       0:00:56.359 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:8
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.115)       0:00:56.474 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1)  => {
    "_storage_test_pool_member_path": "/dev/nvme1n1",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:15
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.027)       0:00:56.502 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/nvme1n1)

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:1
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.042)       0:00:56.544 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:6
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.047)       0:00:56.592 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:11
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.048)       0:00:56.640 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:17
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.036)       0:00:56.677 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:23
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.039)       0:00:56.717 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:29
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.034)       0:00:56.751 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:22
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.032)       0:00:56.784 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:68
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.037)       0:00:56.822 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml:1
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.046)       0:00:56.868 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [get information about VDO deduplication] *********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.044)       0:00:56.913 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.026)       0:00:56.939 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.024)       0:00:56.964 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.024)       0:00:56.989 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.026)       0:00:57.015 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.023)       0:00:57.039 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.026)       0:00:57.065 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.026)       0:00:57.092 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:71
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.039)       0:00:57.132 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [verify the volumes] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml:3
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.033)       0:00:57.165 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2
Thursday 21 July 2022  15:40:16 +0000 (0:00:00.044)       0:00:57.210 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.051)       0:00:57.261 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache)

TASK [Get expected mount device based on device type] **************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.078)       0:00:57.340 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.041)       0:00:57.381 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": [
            {
                "block_available": 938726,
                "block_size": 4096,
                "block_total": 999344,
                "block_used": 60618,
                "device": "/dev/mapper/foo-test1",
                "fstype": "ext4",
                "inode_available": 262133,
                "inode_total": 262144,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 3845021696,
                "size_total": 4093313024,
                "uuid": "f7dcf85c-126f-4bdd-83f1-93b5c5721fc4"
            }
        ],
        "storage_test_mount_expected_match_count": "1",
        "storage_test_mount_point_matches": [
            {
                "block_available": 938726,
                "block_size": 4096,
                "block_total": 999344,
                "block_used": 60618,
                "device": "/dev/mapper/foo-test1",
                "fstype": "ext4",
                "inode_available": 262133,
                "inode_total": 262144,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 3845021696,
                "size_total": 4093313024,
                "uuid": "f7dcf85c-126f-4bdd-83f1-93b5c5721fc4"
            }
        ],
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.057)       0:00:57.439 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the current mount state by mount point] ***************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.053)       0:00:57.492 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the mount fs type] ************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.049)       0:00:57.542 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [command] *****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.055)       0:00:57.597 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.030)       0:00:57.628 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.028)       0:00:57.657 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.024)       0:00:57.682 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": null,
        "storage_test_mount_expected_match_count": null,
        "storage_test_mount_point_matches": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.088)       0:00:57.770 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.127)       0:00:57.898 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.049)       0:00:57.948 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.048)       0:00:57.996 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up variables] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.038)       0:00:58.034 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.034)       0:00:58.069 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.041)       0:00:58.110 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4
Thursday 21 July 2022  15:40:17 +0000 (0:00:00.042)       0:00:58.153 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658418010.5750651,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1658418010.5750651,
        "dev": 5,
        "device_type": 64512,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 40802,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1658418010.5750651,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10
Thursday 21 July 2022  15:40:18 +0000 (0:00:00.329)       0:00:58.482 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18
Thursday 21 July 2022  15:40:18 +0000 (0:00:00.038)       0:00:58.521 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [(1/2) Process volume type (set initial value)] ***************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24
Thursday 21 July 2022  15:40:18 +0000 (0:00:00.043)       0:00:58.564 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [(2/2) Process volume type (get RAID value)] ******************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28
Thursday 21 July 2022  15:40:18 +0000 (0:00:00.040)       0:00:58.604 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33
Thursday 21 July 2022  15:40:18 +0000 (0:00:00.027)       0:00:58.632 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3
Thursday 21 July 2022  15:40:18 +0000 (0:00:00.039)       0:00:58.672 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10
Thursday 21 July 2022  15:40:18 +0000 (0:00:00.025)       0:00:58.698 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "rc": 0,
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.541)       0:00:59.240 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.024)       0:00:59.264 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.026)       0:00:59.290 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.052)       0:00:59.342 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.023)       0:00:59.366 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.023)       0:00:59.389 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.025)       0:00:59.415 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.025)       0:00:59.440 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.023)       0:00:59.464 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.051)       0:00:59.515 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.056)       0:00:59.572 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.039)       0:00:59.611 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.038)       0:00:59.649 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.036)       0:00:59.686 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [get information about RAID] **********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.036)       0:00:59.722 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.035)       0:00:59.758 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.035)       0:00:59.793 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.035)       0:00:59.828 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.039)       0:00:59.868 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.039)       0:00:59.908 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.037)       0:00:59.945 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the actual size of the volume] *************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3
Thursday 21 July 2022  15:40:19 +0000 (0:00:00.035)       0:00:59.981 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [parse the requested size of the volume] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9
Thursday 21 July 2022  15:40:20 +0000 (0:00:00.310)       0:01:00.292 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15
Thursday 21 July 2022  15:40:20 +0000 (0:00:00.302)       0:01:00.594 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_expected_size": "4294967296"
    },
    "changed": false
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20
Thursday 21 July 2022  15:40:20 +0000 (0:00:00.078)       0:01:00.672 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25
Thursday 21 July 2022  15:40:20 +0000 (0:00:00.035)       0:01:00.707 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28
Thursday 21 July 2022  15:40:20 +0000 (0:00:00.040)       0:01:00.748 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31
Thursday 21 July 2022  15:40:20 +0000 (0:00:00.038)       0:01:00.786 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36
Thursday 21 July 2022  15:40:20 +0000 (0:00:00.038)       0:01:00.824 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39
Thursday 21 July 2022  15:40:20 +0000 (0:00:00.037)       0:01:00.861 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44
Thursday 21 July 2022  15:40:20 +0000 (0:00:00.040)       0:01:00.902 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_actual_size": {
        "bytes": 4294967296,
        "changed": false,
        "failed": false,
        "lvm": "4g",
        "parted": "4GiB",
        "size": "4 GiB"
    }
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47
Thursday 21 July 2022  15:40:20 +0000 (0:00:00.037)       0:01:00.939 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50
Thursday 21 July 2022  15:40:20 +0000 (0:00:00.091)       0:01:01.030 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6
Thursday 21 July 2022  15:40:20 +0000 (0:00:00.053)       0:01:01.084 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.031705",
    "end": "2022-07-21 11:40:21.175008",
    "rc": 0,
    "start": "2022-07-21 11:40:21.143303"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14
Thursday 21 July 2022  15:40:21 +0000 (0:00:00.337)       0:01:01.422 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [check segment type] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17
Thursday 21 July 2022  15:40:21 +0000 (0:00:00.089)       0:01:01.511 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22
Thursday 21 July 2022  15:40:21 +0000 (0:00:00.126)       0:01:01.638 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested cache size] ******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26
Thursday 21 July 2022  15:40:21 +0000 (0:00:00.047)       0:01:01.686 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32
Thursday 21 July 2022  15:40:21 +0000 (0:00:00.041)       0:01:01.727 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36
Thursday 21 July 2022  15:40:21 +0000 (0:00:00.037)       0:01:01.764 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16
Thursday 21 July 2022  15:40:21 +0000 (0:00:00.034)       0:01:01.799 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43
Thursday 21 July 2022  15:40:21 +0000 (0:00:00.035)       0:01:01.835 ********* 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53
Thursday 21 July 2022  15:40:21 +0000 (0:00:00.020)       0:01:01.855 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Resizing with one large value which large than disk's size] **************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:123
Thursday 21 July 2022  15:40:21 +0000 (0:00:00.035)       0:01:01.890 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:40:21 +0000 (0:00:00.040)       0:01:01.930 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:40:21 +0000 (0:00:00.036)       0:01:01.967 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:40:22 +0000 (0:00:00.437)       0:01:02.405 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:40:22 +0000 (0:00:00.068)       0:01:02.473 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:40:22 +0000 (0:00:00.033)       0:01:02.507 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:40:22 +0000 (0:00:00.031)       0:01:02.538 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:40:22 +0000 (0:00:00.046)       0:01:02.585 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:40:22 +0000 (0:00:00.023)       0:01:02.608 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:40:22 +0000 (0:00:00.031)       0:01:02.639 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "12884901888.0"
                }
            ]
        }
    ]
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:40:22 +0000 (0:00:00.055)       0:01:02.695 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:40:22 +0000 (0:00:00.040)       0:01:02.736 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:40:22 +0000 (0:00:00.028)       0:01:02.764 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:40:22 +0000 (0:00:00.026)       0:01:02.790 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:40:22 +0000 (0:00:00.025)       0:01:02.816 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:40:22 +0000 (0:00:00.029)       0:01:02.846 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:40:22 +0000 (0:00:00.056)       0:01:02.902 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:40:22 +0000 (0:00:00.024)       0:01:02.927 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

volume 'test1' cannot be resized to '12 GiB'

TASK [linux-system-roles.storage : failed message] *****************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99
Thursday 21 July 2022  15:40:23 +0000 (0:00:01.248)       0:01:04.176 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "changed": false
}

MSG:

{'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': False, 'pools': [{'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', 'size': '12884901888.0', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}], 'volumes': [], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': [], 'msg': "volume 'test1' cannot be resized to '12 GiB'", '_ansible_no_log': False}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:40:23 +0000 (0:00:00.043)       0:01:04.219 ********* 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:142
Thursday 21 July 2022  15:40:24 +0000 (0:00:00.024)       0:01:04.244 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the output when resizing with large size] *************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:148
Thursday 21 July 2022  15:40:24 +0000 (0:00:00.081)       0:01:04.325 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Remove the volume group created above] ***********************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:155
Thursday 21 July 2022  15:40:24 +0000 (0:00:00.091)       0:01:04.417 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:40:24 +0000 (0:00:00.033)       0:01:04.450 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:40:24 +0000 (0:00:00.073)       0:01:04.523 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:40:24 +0000 (0:00:00.425)       0:01:04.949 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:40:24 +0000 (0:00:00.059)       0:01:05.008 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:40:24 +0000 (0:00:00.031)       0:01:05.040 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:40:24 +0000 (0:00:00.032)       0:01:05.073 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:40:24 +0000 (0:00:00.045)       0:01:05.118 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:40:24 +0000 (0:00:00.020)       0:01:05.139 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:40:24 +0000 (0:00:00.030)       0:01:05.169 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "state": "absent",
            "type": "lvm"
        }
    ]
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:40:24 +0000 (0:00:00.038)       0:01:05.207 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:40:25 +0000 (0:00:00.037)       0:01:05.245 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:40:25 +0000 (0:00:00.038)       0:01:05.283 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:40:25 +0000 (0:00:00.029)       0:01:05.313 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:40:25 +0000 (0:00:00.027)       0:01:05.341 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:40:25 +0000 (0:00:00.030)       0:01:05.372 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:40:25 +0000 (0:00:00.060)       0:01:05.432 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:40:25 +0000 (0:00:00.024)       0:01:05.457 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/nvme1n1",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "absent",
            "type": "lvm",
            "volumes": []
        }
    ],
    "volumes": []
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:40:26 +0000 (0:00:01.542)       0:01:07.000 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:40:26 +0000 (0:00:00.041)       0:01:07.042 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:40:26 +0000 (0:00:00.022)       0:01:07.065 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/nvme1n1",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ],
        "volumes": []
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:40:26 +0000 (0:00:00.040)       0:01:07.105 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ]
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:40:26 +0000 (0:00:00.039)       0:01:07.145 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:40:26 +0000 (0:00:00.040)       0:01:07.185 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:40:27 +0000 (0:00:00.348)       0:01:07.534 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:40:27 +0000 (0:00:00.464)       0:01:07.998 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:40:27 +0000 (0:00:00.042)       0:01:08.041 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:40:28 +0000 (0:00:00.460)       0:01:08.502 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658417969.482065,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:40:28 +0000 (0:00:00.317)       0:01:08.819 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:40:28 +0000 (0:00:00.023)       0:01:08.842 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [Create one partition on one disk] ****************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:165
Thursday 21 July 2022  15:40:29 +0000 (0:00:00.905)       0:01:09.748 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:40:29 +0000 (0:00:00.087)       0:01:09.835 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:40:29 +0000 (0:00:00.035)       0:01:09.871 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:40:30 +0000 (0:00:00.416)       0:01:10.288 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:40:30 +0000 (0:00:00.063)       0:01:10.351 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:40:30 +0000 (0:00:00.037)       0:01:10.388 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:40:30 +0000 (0:00:00.032)       0:01:10.421 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:40:30 +0000 (0:00:00.045)       0:01:10.466 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:40:30 +0000 (0:00:00.023)       0:01:10.490 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:40:30 +0000 (0:00:00.029)       0:01:10.519 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "type": "partition",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "type": "partition"
                }
            ]
        }
    ]
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:40:30 +0000 (0:00:00.040)       0:01:10.560 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:40:30 +0000 (0:00:00.038)       0:01:10.599 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:40:30 +0000 (0:00:00.033)       0:01:10.632 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:40:30 +0000 (0:00:00.027)       0:01:10.659 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:40:30 +0000 (0:00:00.025)       0:01:10.685 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:40:30 +0000 (0:00:00.028)       0:01:10.714 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:40:30 +0000 (0:00:00.060)       0:01:10.774 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:40:30 +0000 (0:00:00.023)       0:01:10.797 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/nvme1n1",
            "fs_type": "disklabel"
        },
        {
            "action": "create device",
            "device": "/dev/nvme1n1p1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/nvme1n1p1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd",
        "/dev/nvme1n1p1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "opts": "defaults",
            "passno": 0,
            "path": "/opt/test1",
            "src": "UUID=bf85bc82-488f-4b59-ad5b-67041b396d30",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs"
    ],
    "pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "partition",
            "volumes": [
                {
                    "_device": "/dev/nvme1n1p1",
                    "_kernel_device": "/dev/nvme1n1p1",
                    "_mount_id": "UUID=bf85bc82-488f-4b59-ad5b-67041b396d30",
                    "_raw_device": "/dev/nvme1n1p1",
                    "_raw_kernel_device": "/dev/nvme1n1p1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": 0,
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "partition",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:40:32 +0000 (0:00:01.575)       0:01:12.373 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:40:32 +0000 (0:00:00.035)       0:01:12.409 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:40:32 +0000 (0:00:00.020)       0:01:12.429 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/nvme1n1",
                "fs_type": "disklabel"
            },
            {
                "action": "create device",
                "device": "/dev/nvme1n1p1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/nvme1n1p1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd",
            "/dev/nvme1n1p1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "opts": "defaults",
                "passno": 0,
                "path": "/opt/test1",
                "src": "UUID=bf85bc82-488f-4b59-ad5b-67041b396d30",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "partition",
                "volumes": [
                    {
                        "_device": "/dev/nvme1n1p1",
                        "_kernel_device": "/dev/nvme1n1p1",
                        "_mount_id": "UUID=bf85bc82-488f-4b59-ad5b-67041b396d30",
                        "_raw_device": "/dev/nvme1n1p1",
                        "_raw_kernel_device": "/dev/nvme1n1p1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": 0,
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "partition",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:40:32 +0000 (0:00:00.037)       0:01:12.467 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "partition",
                "volumes": [
                    {
                        "_device": "/dev/nvme1n1p1",
                        "_kernel_device": "/dev/nvme1n1p1",
                        "_mount_id": "UUID=bf85bc82-488f-4b59-ad5b-67041b396d30",
                        "_raw_device": "/dev/nvme1n1p1",
                        "_raw_kernel_device": "/dev/nvme1n1p1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": 0,
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "partition",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:40:32 +0000 (0:00:00.040)       0:01:12.508 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:40:32 +0000 (0:00:00.040)       0:01:12.549 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:40:32 +0000 (0:00:00.036)       0:01:12.585 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:40:32 +0000 (0:00:00.480)       0:01:13.066 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': 'UUID=bf85bc82-488f-4b59-ad5b-67041b396d30', 'dump': 0, 'passno': 0, 'fstype': 'ext4', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "opts": "defaults",
        "passno": 0,
        "path": "/opt/test1",
        "src": "UUID=bf85bc82-488f-4b59-ad5b-67041b396d30",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "UUID=bf85bc82-488f-4b59-ad5b-67041b396d30"
}

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:40:33 +0000 (0:00:00.361)       0:01:13.427 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:40:33 +0000 (0:00:00.518)       0:01:13.946 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658417969.482065,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:40:34 +0000 (0:00:00.378)       0:01:14.325 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:40:34 +0000 (0:00:00.024)       0:01:14.349 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:179
Thursday 21 July 2022  15:40:34 +0000 (0:00:00.843)       0:01:15.192 ********* 
included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1
Thursday 21 July 2022  15:40:35 +0000 (0:00:00.040)       0:01:15.232 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "partition",
            "volumes": [
                {
                    "_device": "/dev/nvme1n1p1",
                    "_kernel_device": "/dev/nvme1n1p1",
                    "_mount_id": "UUID=bf85bc82-488f-4b59-ad5b-67041b396d30",
                    "_raw_device": "/dev/nvme1n1p1",
                    "_raw_kernel_device": "/dev/nvme1n1p1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": 0,
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "partition",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6
Thursday 21 July 2022  15:40:35 +0000 (0:00:00.052)       0:01:15.284 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14
Thursday 21 July 2022  15:40:35 +0000 (0:00:00.037)       0:01:15.322 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1p1": {
            "fstype": "ext4",
            "label": "",
            "name": "/dev/nvme1n1p1",
            "size": "10G",
            "type": "partition",
            "uuid": "bf85bc82-488f-4b59-ad5b-67041b396d30"
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-15-39-07-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19
Thursday 21 July 2022  15:40:35 +0000 (0:00:00.317)       0:01:15.639 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003359",
    "end": "2022-07-21 11:40:35.704226",
    "rc": 0,
    "start": "2022-07-21 11:40:35.700867"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0
UUID=bf85bc82-488f-4b59-ad5b-67041b396d30 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24
Thursday 21 July 2022  15:40:35 +0000 (0:00:00.308)       0:01:15.948 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003421",
    "end": "2022-07-21 11:40:36.013046",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 11:40:36.009625"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.309)       0:01:16.258 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/nvme1n1p1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/nvme1n1p1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/nvme1n1p1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=bf85bc82-488f-4b59-ad5b-67041b396d30', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/nvme1n1p1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'partition', 'encryption_cipher': None, 'raid_spare_count': None})

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:5
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.061)       0:01:16.319 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:18
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.034)       0:01:16.354 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members)
included: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes)

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:1
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.043)       0:01:16.397 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:10
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.024)       0:01:16.422 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:19
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.020)       0:01:16.442 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:23
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.022)       0:01:16.464 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify PV count] *********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:27
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.023)       0:01:16.488 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:34
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.025)       0:01:16.514 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:38
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.023)       0:01:16.538 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:42
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.023)       0:01:16.561 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:46
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.023)       0:01:16.585 ********* 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:56
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.022)       0:01:16.608 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-md.yml for /cache/rhel-7.qcow2

TASK [get information about RAID] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:6
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.067)       0:01:16.675 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:12
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.025)       0:01:16.701 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:16
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.026)       0:01:16.727 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:20
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.022)       0:01:16.750 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:24
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.022)       0:01:16.772 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:30
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.022)       0:01:16.795 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:36
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.025)       0:01:16.820 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:44
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.026)       0:01:16.847 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:59
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.034)       0:01:16.881 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml:1
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.045)       0:01:16.927 ********* 
skipping: [/cache/rhel-7.qcow2] => (item={'_raw_device': '/dev/nvme1n1p1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/nvme1n1p1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/nvme1n1p1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=bf85bc82-488f-4b59-ad5b-67041b396d30', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/nvme1n1p1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})  => {
    "ansible_loop_var": "storage_test_lvmraid_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_lvmraid_volume": {
        "_device": "/dev/nvme1n1p1",
        "_kernel_device": "/dev/nvme1n1p1",
        "_mount_id": "UUID=bf85bc82-488f-4b59-ad5b-67041b396d30",
        "_raw_device": "/dev/nvme1n1p1",
        "_raw_kernel_device": "/dev/nvme1n1p1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [],
        "encryption": false,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": null,
        "encryption_luks_version": null,
        "encryption_password": null,
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "ext4",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "size": 0,
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:62
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.036)       0:01:16.963 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml for /cache/rhel-7.qcow2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml:1
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.044)       0:01:17.008 ********* 
skipping: [/cache/rhel-7.qcow2] => (item={'_raw_device': '/dev/nvme1n1p1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/nvme1n1p1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/nvme1n1p1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=bf85bc82-488f-4b59-ad5b-67041b396d30', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/nvme1n1p1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})  => {
    "ansible_loop_var": "storage_test_thin_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_thin_volume": {
        "_device": "/dev/nvme1n1p1",
        "_kernel_device": "/dev/nvme1n1p1",
        "_mount_id": "UUID=bf85bc82-488f-4b59-ad5b-67041b396d30",
        "_raw_device": "/dev/nvme1n1p1",
        "_raw_kernel_device": "/dev/nvme1n1p1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [],
        "encryption": false,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": null,
        "encryption_luks_version": null,
        "encryption_password": null,
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "ext4",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "size": 0,
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check member encryption] *************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:65
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.030)       0:01:17.039 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:4
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.046)       0:01:17.085 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:8
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.051)       0:01:17.137 ********* 

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:15
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.024)       0:01:17.161 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:22
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.020)       0:01:17.182 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:68
Thursday 21 July 2022  15:40:36 +0000 (0:00:00.034)       0:01:17.217 ********* 
included: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml:1
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.049)       0:01:17.266 ********* 
skipping: [/cache/rhel-7.qcow2] => (item={'_raw_device': '/dev/nvme1n1p1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/nvme1n1p1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/nvme1n1p1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=bf85bc82-488f-4b59-ad5b-67041b396d30', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/nvme1n1p1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})  => {
    "ansible_loop_var": "storage_test_vdo_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_vdo_volume": {
        "_device": "/dev/nvme1n1p1",
        "_kernel_device": "/dev/nvme1n1p1",
        "_mount_id": "UUID=bf85bc82-488f-4b59-ad5b-67041b396d30",
        "_raw_device": "/dev/nvme1n1p1",
        "_raw_kernel_device": "/dev/nvme1n1p1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [],
        "encryption": false,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": null,
        "encryption_luks_version": null,
        "encryption_password": null,
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "ext4",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "size": 0,
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:71
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.031)       0:01:17.298 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [verify the volumes] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml:3
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.037)       0:01:17.336 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/nvme1n1p1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/nvme1n1p1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/nvme1n1p1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=bf85bc82-488f-4b59-ad5b-67041b396d30', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/nvme1n1p1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.043)       0:01:17.379 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.050)       0:01:17.429 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache)

TASK [Get expected mount device based on device type] **************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.083)       0:01:17.513 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/nvme1n1p1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.041)       0:01:17.555 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": [
            {
                "block_available": 2402892,
                "block_size": 4096,
                "block_total": 2547269,
                "block_used": 144377,
                "device": "/dev/nvme1n1p1",
                "fstype": "ext4",
                "inode_available": 655349,
                "inode_total": 655360,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 9842245632,
                "size_total": 10433613824,
                "uuid": "bf85bc82-488f-4b59-ad5b-67041b396d30"
            }
        ],
        "storage_test_mount_expected_match_count": "1",
        "storage_test_mount_point_matches": [
            {
                "block_available": 2402892,
                "block_size": 4096,
                "block_total": 2547269,
                "block_used": 144377,
                "device": "/dev/nvme1n1p1",
                "fstype": "ext4",
                "inode_available": 655349,
                "inode_total": 655360,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 9842245632,
                "size_total": 10433613824,
                "uuid": "bf85bc82-488f-4b59-ad5b-67041b396d30"
            }
        ],
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.058)       0:01:17.614 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the current mount state by mount point] ***************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.055)       0:01:17.669 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the mount fs type] ************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.050)       0:01:17.720 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [command] *****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.053)       0:01:17.773 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.022)       0:01:17.796 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.022)       0:01:17.818 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.023)       0:01:17.842 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": null,
        "storage_test_mount_expected_match_count": null,
        "storage_test_mount_point_matches": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.079)       0:01:17.921 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "UUID=bf85bc82-488f-4b59-ad5b-67041b396d30 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.062)       0:01:17.984 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.049)       0:01:18.033 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.052)       0:01:18.086 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up variables] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.037)       0:01:18.123 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.037)       0:01:18.160 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10
Thursday 21 July 2022  15:40:37 +0000 (0:00:00.042)       0:01:18.203 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4
Thursday 21 July 2022  15:40:38 +0000 (0:00:00.040)       0:01:18.243 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658418032.080065,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1658418032.080065,
        "dev": 5,
        "device_type": 66307,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 51514,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/blockdevice",
        "mode": "0660",
        "mtime": 1658418032.080065,
        "nlink": 1,
        "path": "/dev/nvme1n1p1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10
Thursday 21 July 2022  15:40:38 +0000 (0:00:00.322)       0:01:18.566 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18
Thursday 21 July 2022  15:40:38 +0000 (0:00:00.041)       0:01:18.607 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [(1/2) Process volume type (set initial value)] ***************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24
Thursday 21 July 2022  15:40:38 +0000 (0:00:00.041)       0:01:18.649 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "st_volume_type": "partition"
    },
    "changed": false
}

TASK [(2/2) Process volume type (get RAID value)] ******************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28
Thursday 21 July 2022  15:40:38 +0000 (0:00:00.039)       0:01:18.688 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33
Thursday 21 July 2022  15:40:38 +0000 (0:00:00.025)       0:01:18.713 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3
Thursday 21 July 2022  15:40:38 +0000 (0:00:00.039)       0:01:18.752 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10
Thursday 21 July 2022  15:40:38 +0000 (0:00:00.023)       0:01:18.776 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "rc": 0,
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.547)       0:01:19.323 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.024)       0:01:19.347 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.022)       0:01:19.370 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.050)       0:01:19.420 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.022)       0:01:19.443 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.025)       0:01:19.469 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.022)       0:01:19.492 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.024)       0:01:19.516 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.025)       0:01:19.541 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.054)       0:01:19.595 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.050)       0:01:19.646 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.035)       0:01:19.682 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.038)       0:01:19.720 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.036)       0:01:19.756 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [get information about RAID] **********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.035)       0:01:19.792 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.034)       0:01:19.827 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.037)       0:01:19.864 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.033)       0:01:19.898 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.033)       0:01:19.931 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.036)       0:01:19.968 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.038)       0:01:20.006 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the actual size of the volume] *************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.033)       0:01:20.040 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested size of the volume] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.029)       0:01:20.069 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.038)       0:01:20.108 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.042)       0:01:20.150 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25
Thursday 21 July 2022  15:40:39 +0000 (0:00:00.036)       0:01:20.187 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.037)       0:01:20.224 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.038)       0:01:20.263 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.043)       0:01:20.306 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.038)       0:01:20.344 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.036)       0:01:20.381 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.078)       0:01:20.459 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.078)       0:01:20.538 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.026)       0:01:20.564 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.025)       0:01:20.590 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check segment type] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.024)       0:01:20.615 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.024)       0:01:20.640 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested cache size] ******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.025)       0:01:20.665 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.025)       0:01:20.690 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.026)       0:01:20.716 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.025)       0:01:20.742 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.036)       0:01:20.778 ********* 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.022)       0:01:20.801 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Test setting up disk volume will remove the partition create above] ******
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:181
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.039)       0:01:20.841 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.051)       0:01:20.892 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:40:40 +0000 (0:00:00.036)       0:01:20.928 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:40:41 +0000 (0:00:00.405)       0:01:21.334 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:40:41 +0000 (0:00:00.064)       0:01:21.398 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:40:41 +0000 (0:00:00.034)       0:01:21.432 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:40:41 +0000 (0:00:00.034)       0:01:21.466 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:40:41 +0000 (0:00:00.048)       0:01:21.515 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:40:41 +0000 (0:00:00.021)       0:01:21.536 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:40:41 +0000 (0:00:00.030)       0:01:21.567 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:40:41 +0000 (0:00:00.036)       0:01:21.603 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": [
        {
            "disks": [
                "nvme1n1"
            ],
            "fs_create_options": "-F",
            "fs_type": "ext4",
            "mount_options": "rw,noatime,defaults",
            "mount_point": "/opt/test1",
            "name": "foo",
            "type": "disk"
        }
    ]
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:40:41 +0000 (0:00:00.036)       0:01:21.640 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:40:41 +0000 (0:00:00.025)       0:01:21.666 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:40:41 +0000 (0:00:00.025)       0:01:21.692 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:40:41 +0000 (0:00:00.027)       0:01:21.720 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:40:41 +0000 (0:00:00.033)       0:01:21.753 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:40:41 +0000 (0:00:00.059)       0:01:21.813 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:40:41 +0000 (0:00:00.023)       0:01:21.836 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/nvme1n1p1",
            "fs_type": "ext4"
        },
        {
            "action": "destroy device",
            "device": "/dev/nvme1n1p1",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/nvme1n1",
            "fs_type": "disklabel"
        },
        {
            "action": "create format",
            "device": "/dev/nvme1n1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "UUID=bf85bc82-488f-4b59-ad5b-67041b396d30",
            "state": "absent"
        },
        {
            "dump": 0,
            "fstype": "ext4",
            "opts": "rw,noatime,defaults",
            "passno": 0,
            "path": "/opt/test1",
            "src": "UUID=63e7b667-280e-4582-aea7-aaca11c34482",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs"
    ],
    "pools": [],
    "volumes": [
        {
            "_device": "/dev/nvme1n1",
            "_kernel_device": "/dev/nvme1n1",
            "_mount_id": "UUID=63e7b667-280e-4582-aea7-aaca11c34482",
            "_raw_device": "/dev/nvme1n1",
            "_raw_kernel_device": "/dev/nvme1n1",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "fs_create_options": "-F",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "ext4",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_options": "rw,noatime,defaults",
            "mount_passno": 0,
            "mount_point": "/opt/test1",
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "size": 10737418240,
            "state": "present",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:40:43 +0000 (0:00:01.564)       0:01:23.400 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:40:43 +0000 (0:00:00.042)       0:01:23.443 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:40:43 +0000 (0:00:00.023)       0:01:23.467 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/nvme1n1p1",
                "fs_type": "ext4"
            },
            {
                "action": "destroy device",
                "device": "/dev/nvme1n1p1",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/nvme1n1",
                "fs_type": "disklabel"
            },
            {
                "action": "create format",
                "device": "/dev/nvme1n1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "UUID=bf85bc82-488f-4b59-ad5b-67041b396d30",
                "state": "absent"
            },
            {
                "dump": 0,
                "fstype": "ext4",
                "opts": "rw,noatime,defaults",
                "passno": 0,
                "path": "/opt/test1",
                "src": "UUID=63e7b667-280e-4582-aea7-aaca11c34482",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs"
        ],
        "pools": [],
        "volumes": [
            {
                "_device": "/dev/nvme1n1",
                "_kernel_device": "/dev/nvme1n1",
                "_mount_id": "UUID=63e7b667-280e-4582-aea7-aaca11c34482",
                "_raw_device": "/dev/nvme1n1",
                "_raw_kernel_device": "/dev/nvme1n1",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "-F",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "ext4",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_options": "rw,noatime,defaults",
                "mount_passno": 0,
                "mount_point": "/opt/test1",
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "size": 10737418240,
                "state": "present",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:40:43 +0000 (0:00:00.041)       0:01:23.508 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:40:43 +0000 (0:00:00.038)       0:01:23.547 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": [
            {
                "_device": "/dev/nvme1n1",
                "_kernel_device": "/dev/nvme1n1",
                "_mount_id": "UUID=63e7b667-280e-4582-aea7-aaca11c34482",
                "_raw_device": "/dev/nvme1n1",
                "_raw_kernel_device": "/dev/nvme1n1",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "-F",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "ext4",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_options": "rw,noatime,defaults",
                "mount_passno": 0,
                "mount_point": "/opt/test1",
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "size": 10737418240,
                "state": "present",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:40:43 +0000 (0:00:00.038)       0:01:23.586 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': 'UUID=bf85bc82-488f-4b59-ad5b-67041b396d30', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "UUID=bf85bc82-488f-4b59-ad5b-67041b396d30",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "UUID=bf85bc82-488f-4b59-ad5b-67041b396d30"
}

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:40:43 +0000 (0:00:00.352)       0:01:23.939 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:40:44 +0000 (0:00:00.453)       0:01:24.392 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': 'UUID=63e7b667-280e-4582-aea7-aaca11c34482', 'dump': 0, 'passno': 0, 'fstype': 'ext4', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'rw,noatime,defaults'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "opts": "rw,noatime,defaults",
        "passno": 0,
        "path": "/opt/test1",
        "src": "UUID=63e7b667-280e-4582-aea7-aaca11c34482",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "rw,noatime,defaults",
    "passno": "0",
    "src": "UUID=63e7b667-280e-4582-aea7-aaca11c34482"
}

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:40:44 +0000 (0:00:00.356)       0:01:24.749 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:40:44 +0000 (0:00:00.461)       0:01:25.210 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658417969.482065,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:40:45 +0000 (0:00:00.331)       0:01:25.542 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:40:45 +0000 (0:00:00.022)       0:01:25.564 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:194
Thursday 21 July 2022  15:40:46 +0000 (0:00:00.875)       0:01:26.440 ********* 
included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1
Thursday 21 July 2022  15:40:46 +0000 (0:00:00.041)       0:01:26.482 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Print out volume information] ********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6
Thursday 21 July 2022  15:40:46 +0000 (0:00:00.037)       0:01:26.519 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_volumes_list": [
        {
            "_device": "/dev/nvme1n1",
            "_kernel_device": "/dev/nvme1n1",
            "_mount_id": "UUID=63e7b667-280e-4582-aea7-aaca11c34482",
            "_raw_device": "/dev/nvme1n1",
            "_raw_kernel_device": "/dev/nvme1n1",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "fs_create_options": "-F",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "ext4",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_options": "rw,noatime,defaults",
            "mount_passno": 0,
            "mount_point": "/opt/test1",
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "size": 10737418240,
            "state": "present",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14
Thursday 21 July 2022  15:40:46 +0000 (0:00:00.089)       0:01:26.609 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "ext4",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": "63e7b667-280e-4582-aea7-aaca11c34482"
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-15-39-07-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19
Thursday 21 July 2022  15:40:46 +0000 (0:00:00.342)       0:01:26.952 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003517",
    "end": "2022-07-21 11:40:47.062063",
    "rc": 0,
    "start": "2022-07-21 11:40:47.058546"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0
UUID=63e7b667-280e-4582-aea7-aaca11c34482 /opt/test1 ext4 rw,noatime,defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24
Thursday 21 July 2022  15:40:47 +0000 (0:00:00.354)       0:01:27.306 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003406",
    "end": "2022-07-21 11:40:47.381893",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 11:40:47.378487"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33
Thursday 21 July 2022  15:40:47 +0000 (0:00:00.322)       0:01:27.628 ********* 

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43
Thursday 21 July 2022  15:40:47 +0000 (0:00:00.020)       0:01:27.649 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/nvme1n1', 'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'ext4', 'mount_options': 'rw,noatime,defaults', '_device': '/dev/nvme1n1', 'size': 10737418240, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/nvme1n1', 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'type': 'disk', 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=63e7b667-280e-4582-aea7-aaca11c34482', 'raid_spare_count': None, 'name': 'foo', '_raw_kernel_device': '/dev/nvme1n1', 'cache_mode': None, 'cache_devices': [], 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': None, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'thin_pool_size': None, 'fs_create_options': '-F'})

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2
Thursday 21 July 2022  15:40:47 +0000 (0:00:00.061)       0:01:27.710 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10
Thursday 21 July 2022  15:40:47 +0000 (0:00:00.051)       0:01:27.762 ********* 
included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size)
included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache)

TASK [Get expected mount device based on device type] **************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6
Thursday 21 July 2022  15:40:47 +0000 (0:00:00.074)       0:01:27.837 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/nvme1n1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14
Thursday 21 July 2022  15:40:47 +0000 (0:00:00.039)       0:01:27.877 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": [
            {
                "block_available": 2403135,
                "block_size": 4096,
                "block_total": 2547525,
                "block_used": 144390,
                "device": "/dev/nvme1n1",
                "fstype": "ext4",
                "inode_available": 655349,
                "inode_total": 655360,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,noatime,data=ordered",
                "size_available": 9843240960,
                "size_total": 10434662400,
                "uuid": "63e7b667-280e-4582-aea7-aaca11c34482"
            }
        ],
        "storage_test_mount_expected_match_count": "1",
        "storage_test_mount_point_matches": [
            {
                "block_available": 2403135,
                "block_size": 4096,
                "block_total": 2547525,
                "block_used": 144390,
                "device": "/dev/nvme1n1",
                "fstype": "ext4",
                "inode_available": 655349,
                "inode_total": 655360,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,noatime,data=ordered",
                "size_available": 9843240960,
                "size_total": 10434662400,
                "uuid": "63e7b667-280e-4582-aea7-aaca11c34482"
            }
        ],
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28
Thursday 21 July 2022  15:40:47 +0000 (0:00:00.052)       0:01:27.929 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the current mount state by mount point] ***************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37
Thursday 21 July 2022  15:40:47 +0000 (0:00:00.050)       0:01:27.979 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the mount fs type] ************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45
Thursday 21 July 2022  15:40:47 +0000 (0:00:00.052)       0:01:28.032 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [command] *****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54
Thursday 21 July 2022  15:40:47 +0000 (0:00:00.048)       0:01:28.080 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58
Thursday 21 July 2022  15:40:47 +0000 (0:00:00.024)       0:01:28.105 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63
Thursday 21 July 2022  15:40:47 +0000 (0:00:00.023)       0:01:28.128 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75
Thursday 21 July 2022  15:40:47 +0000 (0:00:00.023)       0:01:28.152 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": null,
        "storage_test_mount_expected_match_count": null,
        "storage_test_mount_point_matches": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2
Thursday 21 July 2022  15:40:47 +0000 (0:00:00.034)       0:01:28.187 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "UUID=63e7b667-280e-4582-aea7-aaca11c34482 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 rw,noatime,defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25
Thursday 21 July 2022  15:40:48 +0000 (0:00:00.062)       0:01:28.249 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32
Thursday 21 July 2022  15:40:48 +0000 (0:00:00.051)       0:01:28.301 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39
Thursday 21 July 2022  15:40:48 +0000 (0:00:00.049)       0:01:28.350 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49
Thursday 21 July 2022  15:40:48 +0000 (0:00:00.052)       0:01:28.402 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4
Thursday 21 July 2022  15:40:48 +0000 (0:00:00.036)       0:01:28.438 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10
Thursday 21 July 2022  15:40:48 +0000 (0:00:00.041)       0:01:28.480 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4
Thursday 21 July 2022  15:40:48 +0000 (0:00:00.039)       0:01:28.519 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658418043.113065,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1658418043.113065,
        "dev": 5,
        "device_type": 66305,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 1806,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/blockdevice",
        "mode": "0660",
        "mtime": 1658418043.113065,
        "nlink": 1,
        "path": "/dev/nvme1n1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10
Thursday 21 July 2022  15:40:48 +0000 (0:00:00.321)       0:01:28.841 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18
Thursday 21 July 2022  15:40:48 +0000 (0:00:00.040)       0:01:28.882 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [(1/2) Process volume type (set initial value)] ***************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24
Thursday 21 July 2022  15:40:48 +0000 (0:00:00.041)       0:01:28.923 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "st_volume_type": "disk"
    },
    "changed": false
}

TASK [(2/2) Process volume type (get RAID value)] ******************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28
Thursday 21 July 2022  15:40:48 +0000 (0:00:00.035)       0:01:28.958 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33
Thursday 21 July 2022  15:40:48 +0000 (0:00:00.029)       0:01:28.988 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3
Thursday 21 July 2022  15:40:48 +0000 (0:00:00.084)       0:01:29.072 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10
Thursday 21 July 2022  15:40:48 +0000 (0:00:00.026)       0:01:29.098 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "rc": 0,
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.530)       0:01:29.629 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.025)       0:01:29.654 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.024)       0:01:29.679 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.052)       0:01:29.731 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.023)       0:01:29.754 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.025)       0:01:29.780 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.024)       0:01:29.804 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.024)       0:01:29.829 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.024)       0:01:29.853 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.055)       0:01:29.908 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.050)       0:01:29.959 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.035)       0:01:29.994 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.034)       0:01:30.029 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.041)       0:01:30.070 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [get information about RAID] **********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.032)       0:01:30.102 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.034)       0:01:30.137 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.037)       0:01:30.174 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21
Thursday 21 July 2022  15:40:49 +0000 (0:00:00.044)       0:01:30.218 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.038)       0:01:30.257 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.037)       0:01:30.294 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.037)       0:01:30.332 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the actual size of the volume] *************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.041)       0:01:30.374 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested size of the volume] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.023)       0:01:30.398 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.038)       0:01:30.436 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.037)       0:01:30.474 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.039)       0:01:30.513 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.038)       0:01:30.551 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.037)       0:01:30.588 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.037)       0:01:30.626 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.040)       0:01:30.667 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.036)       0:01:30.704 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [debug] *******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.037)       0:01:30.741 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [assert] ******************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.038)       0:01:30.780 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.024)       0:01:30.804 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.022)       0:01:30.826 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check segment type] ******************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.022)       0:01:30.848 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.022)       0:01:30.871 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested cache size] ******************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.024)       0:01:30.895 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.022)       0:01:30.918 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.022)       0:01:30.940 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.021)       0:01:30.961 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.104)       0:01:31.066 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Remove the disk volume created above] ************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:198
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.033)       0:01:31.099 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.053)       0:01:31.153 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:40:50 +0000 (0:00:00.034)       0:01:31.187 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:40:51 +0000 (0:00:00.421)       0:01:31.609 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:40:51 +0000 (0:00:00.063)       0:01:31.672 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:40:51 +0000 (0:00:00.036)       0:01:31.708 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:40:51 +0000 (0:00:00.036)       0:01:31.745 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:40:51 +0000 (0:00:00.057)       0:01:31.802 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:40:51 +0000 (0:00:00.023)       0:01:31.825 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:40:51 +0000 (0:00:00.030)       0:01:31.855 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:40:51 +0000 (0:00:00.041)       0:01:31.897 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "state": "absent",
            "type": "disk"
        }
    ]
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:40:51 +0000 (0:00:00.039)       0:01:31.936 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:40:51 +0000 (0:00:00.029)       0:01:31.966 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:40:51 +0000 (0:00:00.027)       0:01:31.993 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:40:51 +0000 (0:00:00.035)       0:01:32.028 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:40:51 +0000 (0:00:00.031)       0:01:32.060 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:40:51 +0000 (0:00:00.059)       0:01:32.119 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:40:51 +0000 (0:00:00.025)       0:01:32.145 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/nvme1n1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "UUID=63e7b667-280e-4582-aea7-aaca11c34482",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [],
    "volumes": [
        {
            "_device": "/dev/nvme1n1",
            "_mount_id": "UUID=63e7b667-280e-4582-aea7-aaca11c34482",
            "_raw_device": "/dev/nvme1n1",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "fs_create_options": "",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "ext4",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_options": "defaults",
            "mount_passno": 0,
            "mount_point": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "size": 10737418240,
            "state": "absent",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91
Thursday 21 July 2022  15:40:53 +0000 (0:00:01.088)       0:01:33.233 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:40:53 +0000 (0:00:00.038)       0:01:33.271 ********* 

TASK [linux-system-roles.storage : show blivet_output] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109
Thursday 21 July 2022  15:40:53 +0000 (0:00:00.021)       0:01:33.293 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/nvme1n1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "UUID=63e7b667-280e-4582-aea7-aaca11c34482",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [],
        "volumes": [
            {
                "_device": "/dev/nvme1n1",
                "_mount_id": "UUID=63e7b667-280e-4582-aea7-aaca11c34482",
                "_raw_device": "/dev/nvme1n1",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "ext4",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_options": "defaults",
                "mount_passno": 0,
                "mount_point": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "size": 10737418240,
                "state": "absent",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    }
}

TASK [linux-system-roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114
Thursday 21 July 2022  15:40:53 +0000 (0:00:00.047)       0:01:33.340 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118
Thursday 21 July 2022  15:40:53 +0000 (0:00:00.087)       0:01:33.427 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": [
            {
                "_device": "/dev/nvme1n1",
                "_mount_id": "UUID=63e7b667-280e-4582-aea7-aaca11c34482",
                "_raw_device": "/dev/nvme1n1",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "ext4",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_options": "defaults",
                "mount_passno": 0,
                "mount_point": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "size": 10737418240,
                "state": "absent",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    },
    "changed": false
}

TASK [linux-system-roles.storage : remove obsolete mounts] *********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134
Thursday 21 July 2022  15:40:53 +0000 (0:00:00.038)       0:01:33.465 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': 'UUID=63e7b667-280e-4582-aea7-aaca11c34482', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "UUID=63e7b667-280e-4582-aea7-aaca11c34482",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "UUID=63e7b667-280e-4582-aea7-aaca11c34482"
}

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146
Thursday 21 July 2022  15:40:53 +0000 (0:00:00.438)       0:01:33.904 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : set up new/current mounts] ******************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151
Thursday 21 July 2022  15:40:54 +0000 (0:00:00.493)       0:01:34.397 ********* 

TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163
Thursday 21 July 2022  15:40:54 +0000 (0:00:00.036)       0:01:34.434 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171
Thursday 21 July 2022  15:40:54 +0000 (0:00:00.442)       0:01:34.877 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658417969.482065,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176
Thursday 21 July 2022  15:40:54 +0000 (0:00:00.321)       0:01:35.198 ********* 

TASK [linux-system-roles.storage : Update facts] *******************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198
Thursday 21 July 2022  15:40:55 +0000 (0:00:00.021)       0:01:35.220 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [Try to mount swap filesystem to "/opt/test1"] ****************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:210
Thursday 21 July 2022  15:40:55 +0000 (0:00:00.835)       0:01:36.055 ********* 

TASK [linux-system-roles.storage : set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2
Thursday 21 July 2022  15:40:55 +0000 (0:00:00.042)       0:01:36.097 ********* 
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : Ensure ansible_facts used by role] **********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2
Thursday 21 July 2022  15:40:55 +0000 (0:00:00.034)       0:01:36.131 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [linux-system-roles.storage : Set platform/version specific variables] ****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8
Thursday 21 July 2022  15:40:56 +0000 (0:00:00.426)       0:01:36.558 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5
Thursday 21 July 2022  15:40:56 +0000 (0:00:00.064)       0:01:36.622 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9
Thursday 21 July 2022  15:40:56 +0000 (0:00:00.036)       0:01:36.659 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : include the appropriate provider tasks] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13
Thursday 21 July 2022  15:40:56 +0000 (0:00:00.033)       0:01:36.693 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  15:40:56 +0000 (0:00:00.048)       0:01:36.741 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure blivet is available] **************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  15:40:56 +0000 (0:00:00.022)       0:01:36.764 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : show storage_pools] *************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14
Thursday 21 July 2022  15:40:56 +0000 (0:00:00.028)       0:01:36.792 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [linux-system-roles.storage : show storage_volumes] ***********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19
Thursday 21 July 2022  15:40:56 +0000 (0:00:00.074)       0:01:36.866 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": [
        {
            "disks": [
                "nvme1n1"
            ],
            "fs_type": "swap",
            "mount_point": "/opt/test1",
            "name": "test1",
            "type": "disk"
        }
    ]
}

TASK [linux-system-roles.storage : get required packages] **********************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24
Thursday 21 July 2022  15:40:56 +0000 (0:00:00.038)       0:01:36.904 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : enable copr repositories if needed] *********
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37
Thursday 21 July 2022  15:40:56 +0000 (0:00:00.025)       0:01:36.930 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : make sure required packages are installed] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44
Thursday 21 July 2022  15:40:56 +0000 (0:00:00.025)       0:01:36.956 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : get service facts] **************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51
Thursday 21 July 2022  15:40:56 +0000 (0:00:00.029)       0:01:36.985 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  15:40:56 +0000 (0:00:00.027)       0:01:37.013 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] *******
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71
Thursday 21 July 2022  15:40:56 +0000 (0:00:00.054)       0:01:37.068 ********* 

TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77
Thursday 21 July 2022  15:40:56 +0000 (0:00:00.020)       0:01:37.088 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

volume 'test1' has a mount point but no mountable file system

TASK [linux-system-roles.storage : failed message] *****************************
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99
Thursday 21 July 2022  15:40:57 +0000 (0:00:00.915)       0:01:38.004 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "changed": false
}

MSG:

{'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': False, 'pools': [], 'volumes': [{'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'swap', 'mount_options': 'defaults', 'size': 10737418240, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'type': 'disk', 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, 'raid_spare_count': None, 'name': 'test1', 'cache_mode': None, 'cache_devices': [], 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': None, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'thin_pool_size': None, 'fs_create_options': ''}], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': [], 'msg': "volume 'test1' has a mount point but no mountable file system", '_ansible_no_log': False}

TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] *****
task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103
Thursday 21 July 2022  15:40:57 +0000 (0:00:00.039)       0:01:38.043 ********* 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:226
Thursday 21 July 2022  15:40:57 +0000 (0:00:00.020)       0:01:38.063 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the output when mount swap filesystem to "/opt/test1"] ************
task path: /tmp/tmpaxjje44y/tests/tests_misc.yml:232
Thursday 21 July 2022  15:40:57 +0000 (0:00:00.038)       0:01:38.102 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed
META: ran handlers
META: ran handlers

PLAY RECAP *********************************************************************
/cache/rhel-7.qcow2        : ok=481  changed=19   unreachable=0    failed=3    skipped=367  rescued=3    ignored=0   

Thursday 21 July 2022  15:40:57 +0000 (0:00:00.057)       0:01:38.160 ********* 
=============================================================================== 
linux-system-roles.storage : make sure blivet is available -------------- 8.91s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 
linux-system-roles.storage : Update facts ------------------------------- 1.85s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.66s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.61s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.60s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.58s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.56s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.54s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
Ensure cryptsetup is present -------------------------------------------- 1.51s
/tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 -------------------
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.46s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.43s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
set up internal repositories -------------------------------------------- 1.42s
/cache/rhel-7_setup.yml:5 -----------------------------------------------------
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.25s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
linux-system-roles.storage : get service facts -------------------------- 1.13s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.09s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
Gathering Facts --------------------------------------------------------- 1.00s
/tmp/tmpaxjje44y/tests/tests_misc.yml:2 ---------------------------------------
linux-system-roles.storage : manage the pools and volumes to match the specified state --- 0.92s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 
linux-system-roles.storage : Update facts ------------------------------- 0.91s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 
linux-system-roles.storage : Update facts ------------------------------- 0.88s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 
linux-system-roles.storage : Update facts ------------------------------- 0.85s
/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 
ansible-playbook [core 2.12.6]
  config file = /etc/ansible/ansible.cfg
  configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
  ansible python module location = /usr/lib/python3.9/site-packages/ansible
  ansible collection location = /tmp/tmp5bkr4li_
  executable location = /usr/bin/ansible-playbook
  python version = 3.9.13 (main, May 18 2022, 00:00:00) [GCC 11.3.1 20220421 (Red Hat 11.3.1-2)]
  jinja version = 2.11.3
  libyaml = True
Using /etc/ansible/ansible.cfg as config file
Skipping callback 'debug', as we already have a stdout callback.
Skipping callback 'default', as we already have a stdout callback.
Skipping callback 'minimal', as we already have a stdout callback.
Skipping callback 'oneline', as we already have a stdout callback.

PLAYBOOK: rhel-7_setup.yml *****************************************************
1 plays in /cache/rhel-7_setup.yml

PLAY [Setup repos] *************************************************************
META: ran handlers

TASK [set up internal repositories] ********************************************
task path: /cache/rhel-7_setup.yml:5
Thursday 21 July 2022  18:48:44 +0000 (0:00:00.027)       0:00:00.027 ********* 
changed: [/cache/rhel-7.qcow2] => (item=None) => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
changed: [/cache/rhel-7.qcow2] => (item=None) => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
changed: [/cache/rhel-7.qcow2] => (item=None) => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
changed: [/cache/rhel-7.qcow2] => (item=None) => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
changed: [/cache/rhel-7.qcow2] => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
    "changed": true
}
META: ran handlers
META: ran handlers

PLAY RECAP *********************************************************************
/cache/rhel-7.qcow2        : ok=1    changed=1    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   

Thursday 21 July 2022  18:48:45 +0000 (0:00:01.379)       0:00:01.406 ********* 
=============================================================================== 
set up internal repositories -------------------------------------------- 1.38s
/cache/rhel-7_setup.yml:5 -----------------------------------------------------

PLAYBOOK: tests_misc.yml *******************************************************
1 plays in /tmp/tmptomayb7j/tests/storage/tests_misc.yml

PLAY [all] *********************************************************************

TASK [Gathering Facts] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:2
Thursday 21 July 2022  18:48:45 +0000 (0:00:00.020)       0:00:01.427 ********* 
ok: [/cache/rhel-7.qcow2]
META: ran handlers

TASK [include_role : fedora.linux_system_roles.storage] ************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:15
Thursday 21 July 2022  18:48:46 +0000 (0:00:00.978)       0:00:02.406 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:48:46 +0000 (0:00:00.034)       0:00:02.441 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:48:46 +0000 (0:00:00.029)       0:00:02.470 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:48:46 +0000 (0:00:00.393)       0:00:02.864 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:48:47 +0000 (0:00:00.053)       0:00:02.917 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:48:47 +0000 (0:00:00.029)       0:00:02.947 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:48:47 +0000 (0:00:00.030)       0:00:02.977 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:48:47 +0000 (0:00:00.053)       0:00:03.031 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:48:47 +0000 (0:00:00.017)       0:00:03.048 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "changed": true,
    "changes": {
        "installed": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "rc": 0,
    "results": [
        "Loaded plugins: search-disabled-repos\nResolving Dependencies\n--> Running transaction check\n---> Package libblockdev-crypto.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libblockdev-utils(x86-64) = 2.18-5.el7 for package: libblockdev-crypto-2.18-5.el7.x86_64\n--> Processing Dependency: libvolume_key.so.1()(64bit) for package: libblockdev-crypto-2.18-5.el7.x86_64\n--> Processing Dependency: libbd_utils.so.2()(64bit) for package: libblockdev-crypto-2.18-5.el7.x86_64\n---> Package libblockdev-dm.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libdmraid.so.1(Base)(64bit) for package: libblockdev-dm-2.18-5.el7.x86_64\n--> Processing Dependency: dmraid for package: libblockdev-dm-2.18-5.el7.x86_64\n--> Processing Dependency: libdmraid.so.1()(64bit) for package: libblockdev-dm-2.18-5.el7.x86_64\n---> Package libblockdev-lvm.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: lvm2 for package: libblockdev-lvm-2.18-5.el7.x86_64\n--> Processing Dependency: device-mapper-persistent-data for package: libblockdev-lvm-2.18-5.el7.x86_64\n---> Package libblockdev-mdraid.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: mdadm for package: libblockdev-mdraid-2.18-5.el7.x86_64\n--> Processing Dependency: libbytesize.so.1()(64bit) for package: libblockdev-mdraid-2.18-5.el7.x86_64\n---> Package libblockdev-swap.x86_64 0:2.18-5.el7 will be installed\n---> Package python-enum34.noarch 0:1.0.4-1.el7 will be installed\n---> Package python2-blivet3.noarch 1:3.1.3-3.el7 will be installed\n--> Processing Dependency: blivet3-data = 1:3.1.3-3.el7 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-bytesize >= 0.3 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-blockdev >= 2.17 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: pyparted >= 3.9 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-hawkey for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: lsof for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Running transaction check\n---> Package blivet3-data.noarch 1:3.1.3-3.el7 will be installed\n---> Package device-mapper-persistent-data.x86_64 0:0.8.5-3.el7_9.2 will be installed\n--> Processing Dependency: libaio.so.1(LIBAIO_0.4)(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n--> Processing Dependency: libaio.so.1(LIBAIO_0.1)(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n--> Processing Dependency: libaio.so.1()(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n---> Package dmraid.x86_64 0:1.0.0.rc16-28.el7 will be installed\n--> Processing Dependency: libdevmapper-event.so.1.02(Base)(64bit) for package: dmraid-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: dmraid-events for package: dmraid-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: libdevmapper-event.so.1.02()(64bit) for package: dmraid-1.0.0.rc16-28.el7.x86_64\n---> Package libblockdev-utils.x86_64 0:2.18-5.el7 will be installed\n---> Package libbytesize.x86_64 0:1.2-1.el7 will be installed\n--> Processing Dependency: libmpfr.so.4()(64bit) for package: libbytesize-1.2-1.el7.x86_64\n---> Package lsof.x86_64 0:4.87-6.el7 will be installed\n---> Package lvm2.x86_64 7:2.02.187-6.el7_9.5 will be installed\n--> Processing Dependency: lvm2-libs = 7:2.02.187-6.el7_9.5 for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n--> Processing Dependency: liblvm2app.so.2.2(Base)(64bit) for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n--> Processing Dependency: liblvm2app.so.2.2()(64bit) for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n---> Package mdadm.x86_64 0:4.1-9.el7_9 will be installed\n--> Processing Dependency: libreport-filesystem for package: mdadm-4.1-9.el7_9.x86_64\n---> Package pyparted.x86_64 1:3.9-15.el7 will be installed\n---> Package python2-blockdev.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libblockdev(x86-64) = 2.18-5.el7 for package: python2-blockdev-2.18-5.el7.x86_64\n---> Package python2-bytesize.x86_64 0:1.2-1.el7 will be installed\n---> Package python2-hawkey.x86_64 0:0.22.5-2.el7_9 will be installed\n--> Processing Dependency: libdnf(x86-64) = 0.22.5-2.el7_9 for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: python2-libdnf = 0.22.5-2.el7_9 for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolv.so.0(SOLV_1.0)(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolvext.so.0(SOLV_1.0)(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libdnf.so.2()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libjson-glib-1.0.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libmodulemd.so.1()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: librepo.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: librhsm.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolv.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolvext.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n---> Package volume_key-libs.x86_64 0:0.3.9-9.el7 will be installed\n--> Running transaction check\n---> Package device-mapper-event-libs.x86_64 7:1.02.170-6.el7_9.5 will be installed\n---> Package dmraid-events.x86_64 0:1.0.0.rc16-28.el7 will be installed\n--> Processing Dependency: sgpio for package: dmraid-events-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: device-mapper-event for package: dmraid-events-1.0.0.rc16-28.el7.x86_64\n---> Package json-glib.x86_64 0:1.4.2-2.el7 will be installed\n---> Package libaio.x86_64 0:0.3.109-13.el7 will be installed\n---> Package libblockdev.x86_64 0:2.18-5.el7 will be installed\n---> Package libdnf.x86_64 0:0.22.5-2.el7_9 will be installed\n---> Package libmodulemd.x86_64 0:1.6.3-1.el7 will be installed\n---> Package librepo.x86_64 0:1.8.1-8.el7_9 will be installed\n---> Package libreport-filesystem.x86_64 0:2.1.11-53.el7 will be installed\n---> Package librhsm.x86_64 0:0.0.3-3.el7_9 will be installed\n---> Package libsolv.x86_64 0:0.6.34-4.el7 will be installed\n---> Package lvm2-libs.x86_64 7:2.02.187-6.el7_9.5 will be installed\n---> Package mpfr.x86_64 0:3.1.1-4.el7 will be installed\n---> Package python2-libdnf.x86_64 0:0.22.5-2.el7_9 will be installed\n--> Running transaction check\n---> Package device-mapper-event.x86_64 7:1.02.170-6.el7_9.5 will be installed\n---> Package sgpio.x86_64 0:1.2.0.10-13.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package            Arch   Version              Repository                 Size\n================================================================================\nInstalling:\n libblockdev-crypto x86_64 2.18-5.el7           rhel                       60 k\n libblockdev-dm     x86_64 2.18-5.el7           rhel-optional              54 k\n libblockdev-lvm    x86_64 2.18-5.el7           rhel                       62 k\n libblockdev-mdraid x86_64 2.18-5.el7           rhel                       57 k\n libblockdev-swap   x86_64 2.18-5.el7           rhel                       52 k\n python-enum34      noarch 1.0.4-1.el7          rhel                       52 k\n python2-blivet3    noarch 1:3.1.3-3.el7        rhel                      851 k\nInstalling for dependencies:\n blivet3-data       noarch 1:3.1.3-3.el7        rhel                       77 k\n device-mapper-event\n                    x86_64 7:1.02.170-6.el7_9.5 rhel                      192 k\n device-mapper-event-libs\n                    x86_64 7:1.02.170-6.el7_9.5 rhel                      192 k\n device-mapper-persistent-data\n                    x86_64 0.8.5-3.el7_9.2      rhel                      423 k\n dmraid             x86_64 1.0.0.rc16-28.el7    rhel                      151 k\n dmraid-events      x86_64 1.0.0.rc16-28.el7    rhel                       21 k\n json-glib          x86_64 1.4.2-2.el7          rhel                      134 k\n libaio             x86_64 0.3.109-13.el7       rhel                       24 k\n libblockdev        x86_64 2.18-5.el7           rhel                      119 k\n libblockdev-utils  x86_64 2.18-5.el7           rhel                       59 k\n libbytesize        x86_64 1.2-1.el7            rhel                       52 k\n libdnf             x86_64 0.22.5-2.el7_9       rhel-7-server-extras-rpms 536 k\n libmodulemd        x86_64 1.6.3-1.el7          rhel-7-server-extras-rpms 153 k\n librepo            x86_64 1.8.1-8.el7_9        rhel                       82 k\n libreport-filesystem\n                    x86_64 2.1.11-53.el7        rhel                       41 k\n librhsm            x86_64 0.0.3-3.el7_9        rhel-7-server-extras-rpms  28 k\n libsolv            x86_64 0.6.34-4.el7         rhel                      329 k\n lsof               x86_64 4.87-6.el7           rhel                      331 k\n lvm2               x86_64 7:2.02.187-6.el7_9.5 rhel                      1.3 M\n lvm2-libs          x86_64 7:2.02.187-6.el7_9.5 rhel                      1.1 M\n mdadm              x86_64 4.1-9.el7_9          rhel                      440 k\n mpfr               x86_64 3.1.1-4.el7          rhel                      203 k\n pyparted           x86_64 1:3.9-15.el7         rhel                      195 k\n python2-blockdev   x86_64 2.18-5.el7           rhel                       61 k\n python2-bytesize   x86_64 1.2-1.el7            rhel                       22 k\n python2-hawkey     x86_64 0.22.5-2.el7_9       rhel-7-server-extras-rpms  71 k\n python2-libdnf     x86_64 0.22.5-2.el7_9       rhel-7-server-extras-rpms 611 k\n sgpio              x86_64 1.2.0.10-13.el7      rhel                       14 k\n volume_key-libs    x86_64 0.3.9-9.el7          rhel                      141 k\n\nTransaction Summary\n================================================================================\nInstall  7 Packages (+29 Dependent packages)\n\nTotal download size: 8.2 M\nInstalled size: 24 M\nDownloading packages:\n--------------------------------------------------------------------------------\nTotal                                               21 MB/s | 8.2 MB  00:00     \nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n  Installing : libblockdev-utils-2.18-5.el7.x86_64                         1/36 \n  Installing : 7:device-mapper-event-libs-1.02.170-6.el7_9.5.x86_64        2/36 \n  Installing : json-glib-1.4.2-2.el7.x86_64                                3/36 \n  Installing : librhsm-0.0.3-3.el7_9.x86_64                                4/36 \n  Installing : libsolv-0.6.34-4.el7.x86_64                                 5/36 \n  Installing : libaio-0.3.109-13.el7.x86_64                                6/36 \n  Installing : librepo-1.8.1-8.el7_9.x86_64                                7/36 \n  Installing : libmodulemd-1.6.3-1.el7.x86_64                              8/36 \n  Installing : libdnf-0.22.5-2.el7_9.x86_64                                9/36 \n  Installing : device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64       10/36 \n  Installing : 7:device-mapper-event-1.02.170-6.el7_9.5.x86_64            11/36 \n  Installing : 7:lvm2-libs-2.02.187-6.el7_9.5.x86_64                      12/36 \n  Installing : 7:lvm2-2.02.187-6.el7_9.5.x86_64                           13/36 \n  Installing : python2-libdnf-0.22.5-2.el7_9.x86_64                       14/36 \n  Installing : python2-hawkey-0.22.5-2.el7_9.x86_64                       15/36 \n  Installing : libblockdev-2.18-5.el7.x86_64                              16/36 \n  Installing : python2-blockdev-2.18-5.el7.x86_64                         17/36 \n  Installing : 1:pyparted-3.9-15.el7.x86_64                               18/36 \n  Installing : sgpio-1.2.0.10-13.el7.x86_64                               19/36 \n  Installing : dmraid-1.0.0.rc16-28.el7.x86_64                            20/36 \n  Installing : dmraid-events-1.0.0.rc16-28.el7.x86_64                     21/36 \n  Installing : volume_key-libs-0.3.9-9.el7.x86_64                         22/36 \n  Installing : mpfr-3.1.1-4.el7.x86_64                                    23/36 \n  Installing : libbytesize-1.2-1.el7.x86_64                               24/36 \n  Installing : python2-bytesize-1.2-1.el7.x86_64                          25/36 \n  Installing : libreport-filesystem-2.1.11-53.el7.x86_64                  26/36 \n  Installing : mdadm-4.1-9.el7_9.x86_64                                   27/36 \n  Installing : 1:blivet3-data-3.1.3-3.el7.noarch                          28/36 \n  Installing : lsof-4.87-6.el7.x86_64                                     29/36 \n  Installing : 1:python2-blivet3-3.1.3-3.el7.noarch                       30/36 \n  Installing : libblockdev-mdraid-2.18-5.el7.x86_64                       31/36 \n  Installing : libblockdev-crypto-2.18-5.el7.x86_64                       32/36 \n  Installing : libblockdev-dm-2.18-5.el7.x86_64                           33/36 \n  Installing : libblockdev-lvm-2.18-5.el7.x86_64                          34/36 \n  Installing : libblockdev-swap-2.18-5.el7.x86_64                         35/36 \n  Installing : python-enum34-1.0.4-1.el7.noarch                           36/36 \n  Verifying  : 7:device-mapper-event-1.02.170-6.el7_9.5.x86_64             1/36 \n  Verifying  : libblockdev-swap-2.18-5.el7.x86_64                          2/36 \n  Verifying  : librhsm-0.0.3-3.el7_9.x86_64                                3/36 \n  Verifying  : libblockdev-lvm-2.18-5.el7.x86_64                           4/36 \n  Verifying  : lsof-4.87-6.el7.x86_64                                      5/36 \n  Verifying  : libblockdev-mdraid-2.18-5.el7.x86_64                        6/36 \n  Verifying  : libdnf-0.22.5-2.el7_9.x86_64                                7/36 \n  Verifying  : python-enum34-1.0.4-1.el7.noarch                            8/36 \n  Verifying  : 1:blivet3-data-3.1.3-3.el7.noarch                           9/36 \n  Verifying  : dmraid-events-1.0.0.rc16-28.el7.x86_64                     10/36 \n  Verifying  : python2-blockdev-2.18-5.el7.x86_64                         11/36 \n  Verifying  : libmodulemd-1.6.3-1.el7.x86_64                             12/36 \n  Verifying  : librepo-1.8.1-8.el7_9.x86_64                               13/36 \n  Verifying  : libblockdev-dm-2.18-5.el7.x86_64                           14/36 \n  Verifying  : json-glib-1.4.2-2.el7.x86_64                               15/36 \n  Verifying  : libaio-0.3.109-13.el7.x86_64                               16/36 \n  Verifying  : 7:lvm2-libs-2.02.187-6.el7_9.5.x86_64                      17/36 \n  Verifying  : python2-hawkey-0.22.5-2.el7_9.x86_64                       18/36 \n  Verifying  : python2-bytesize-1.2-1.el7.x86_64                          19/36 \n  Verifying  : libblockdev-2.18-5.el7.x86_64                              20/36 \n  Verifying  : libreport-filesystem-2.1.11-53.el7.x86_64                  21/36 \n  Verifying  : libbytesize-1.2-1.el7.x86_64                               22/36 \n  Verifying  : 7:device-mapper-event-libs-1.02.170-6.el7_9.5.x86_64       23/36 \n  Verifying  : python2-libdnf-0.22.5-2.el7_9.x86_64                       24/36 \n  Verifying  : 7:lvm2-2.02.187-6.el7_9.5.x86_64                           25/36 \n  Verifying  : libblockdev-utils-2.18-5.el7.x86_64                        26/36 \n  Verifying  : mpfr-3.1.1-4.el7.x86_64                                    27/36 \n  Verifying  : volume_key-libs-0.3.9-9.el7.x86_64                         28/36 \n  Verifying  : libsolv-0.6.34-4.el7.x86_64                                29/36 \n  Verifying  : device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64       30/36 \n  Verifying  : 1:python2-blivet3-3.1.3-3.el7.noarch                       31/36 \n  Verifying  : dmraid-1.0.0.rc16-28.el7.x86_64                            32/36 \n  Verifying  : mdadm-4.1-9.el7_9.x86_64                                   33/36 \n  Verifying  : sgpio-1.2.0.10-13.el7.x86_64                               34/36 \n  Verifying  : libblockdev-crypto-2.18-5.el7.x86_64                       35/36 \n  Verifying  : 1:pyparted-3.9-15.el7.x86_64                               36/36 \n\nInstalled:\n  libblockdev-crypto.x86_64 0:2.18-5.el7 libblockdev-dm.x86_64 0:2.18-5.el7    \n  libblockdev-lvm.x86_64 0:2.18-5.el7    libblockdev-mdraid.x86_64 0:2.18-5.el7\n  libblockdev-swap.x86_64 0:2.18-5.el7   python-enum34.noarch 0:1.0.4-1.el7    \n  python2-blivet3.noarch 1:3.1.3-3.el7  \n\nDependency Installed:\n  blivet3-data.noarch 1:3.1.3-3.el7                                             \n  device-mapper-event.x86_64 7:1.02.170-6.el7_9.5                               \n  device-mapper-event-libs.x86_64 7:1.02.170-6.el7_9.5                          \n  device-mapper-persistent-data.x86_64 0:0.8.5-3.el7_9.2                        \n  dmraid.x86_64 0:1.0.0.rc16-28.el7                                             \n  dmraid-events.x86_64 0:1.0.0.rc16-28.el7                                      \n  json-glib.x86_64 0:1.4.2-2.el7                                                \n  libaio.x86_64 0:0.3.109-13.el7                                                \n  libblockdev.x86_64 0:2.18-5.el7                                               \n  libblockdev-utils.x86_64 0:2.18-5.el7                                         \n  libbytesize.x86_64 0:1.2-1.el7                                                \n  libdnf.x86_64 0:0.22.5-2.el7_9                                                \n  libmodulemd.x86_64 0:1.6.3-1.el7                                              \n  librepo.x86_64 0:1.8.1-8.el7_9                                                \n  libreport-filesystem.x86_64 0:2.1.11-53.el7                                   \n  librhsm.x86_64 0:0.0.3-3.el7_9                                                \n  libsolv.x86_64 0:0.6.34-4.el7                                                 \n  lsof.x86_64 0:4.87-6.el7                                                      \n  lvm2.x86_64 7:2.02.187-6.el7_9.5                                              \n  lvm2-libs.x86_64 7:2.02.187-6.el7_9.5                                         \n  mdadm.x86_64 0:4.1-9.el7_9                                                    \n  mpfr.x86_64 0:3.1.1-4.el7                                                     \n  pyparted.x86_64 1:3.9-15.el7                                                  \n  python2-blockdev.x86_64 0:2.18-5.el7                                          \n  python2-bytesize.x86_64 0:1.2-1.el7                                           \n  python2-hawkey.x86_64 0:0.22.5-2.el7_9                                        \n  python2-libdnf.x86_64 0:0.22.5-2.el7_9                                        \n  sgpio.x86_64 0:1.2.0.10-13.el7                                                \n  volume_key-libs.x86_64 0:0.3.9-9.el7                                          \n\nComplete!\n"
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:48:55 +0000 (0:00:08.621)       0:00:11.670 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:48:55 +0000 (0:00:00.058)       0:00:11.728 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:48:55 +0000 (0:00:00.031)       0:00:11.760 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:48:56 +0000 (0:00:00.599)       0:00:12.360 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2
Thursday 21 July 2022  18:48:56 +0000 (0:00:00.042)       0:00:12.402 ********* 

TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13
Thursday 21 July 2022  18:48:56 +0000 (0:00:00.030)       0:00:12.433 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable COPRs] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18
Thursday 21 July 2022  18:48:56 +0000 (0:00:00.035)       0:00:12.469 ********* 

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:48:56 +0000 (0:00:00.031)       0:00:12.500 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "rc": 0,
    "results": []
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:48:57 +0000 (0:00:00.534)       0:00:13.034 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "NetworkManager.service": {
                "name": "NetworkManager.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "arp-ethers.service": {
                "name": "arp-ethers.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "auditd.service": {
                "name": "auditd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "autovt@.service": {
                "name": "autovt@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "blivet.service": {
                "name": "blivet.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "blk-availability.service": {
                "name": "blk-availability.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "brandbot.service": {
                "name": "brandbot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "chrony-wait.service": {
                "name": "chrony-wait.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "chronyd.service": {
                "name": "chronyd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "cloud-config.service": {
                "name": "cloud-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-final.service": {
                "name": "cloud-final.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init-local.service": {
                "name": "cloud-init-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "cloud-init.service": {
                "name": "cloud-init.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "console-getty.service": {
                "name": "console-getty.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "console-shell.service": {
                "name": "console-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "container-getty@.service": {
                "name": "container-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "cpupower.service": {
                "name": "cpupower.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "crond.service": {
                "name": "crond.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "dbus.service": {
                "name": "dbus.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "debug-shell.service": {
                "name": "debug-shell.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "display-manager.service": {
                "name": "display-manager.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "dm-event.service": {
                "name": "dm-event.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dmraid-activation.service": {
                "name": "dmraid-activation.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-mount.service": {
                "name": "dracut-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "emergency.service": {
                "name": "emergency.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "exim.service": {
                "name": "exim.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "fstrim.service": {
                "name": "fstrim.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "getty@.service": {
                "name": "getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "enabled"
            },
            "getty@tty1.service": {
                "name": "getty@tty1.service",
                "source": "systemd",
                "state": "running",
                "status": "active"
            },
            "gssproxy.service": {
                "name": "gssproxy.service",
                "source": "systemd",
                "state": "running",
                "status": "disabled"
            },
            "halt-local.service": {
                "name": "halt-local.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "ip6tables.service": {
                "name": "ip6tables.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "iptables.service": {
                "name": "iptables.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "irqbalance.service": {
                "name": "irqbalance.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "kdump.service": {
                "name": "kdump.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-activation-early.service": {
                "name": "lvm2-activation-early.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "lvm2-activation.service": {
                "name": "lvm2-activation.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "lvm2-lvmetad.service": {
                "name": "lvm2-lvmetad.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "lvm2-lvmpolld.service": {
                "name": "lvm2-lvmpolld.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "lvm2-monitor.service": {
                "name": "lvm2-monitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "lvm2-pvscan@.service": {
                "name": "lvm2-pvscan@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-grow-continue@.service": {
                "name": "mdadm-grow-continue@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdadm-last-resort@.service": {
                "name": "mdadm-last-resort@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdcheck_continue.service": {
                "name": "mdcheck_continue.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdcheck_start.service": {
                "name": "mdcheck_start.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmon@.service": {
                "name": "mdmon@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "mdmonitor-oneshot.service": {
                "name": "mdmonitor-oneshot.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "mdmonitor.service": {
                "name": "mdmonitor.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "messagebus.service": {
                "name": "messagebus.service",
                "source": "systemd",
                "state": "active",
                "status": "static"
            },
            "microcode.service": {
                "name": "microcode.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "netconsole": {
                "name": "netconsole",
                "source": "sysv",
                "state": "stopped",
                "status": "disabled"
            },
            "network": {
                "name": "network",
                "source": "sysv",
                "state": "running",
                "status": "enabled"
            },
            "network.service": {
                "name": "network.service",
                "source": "systemd",
                "state": "stopped",
                "status": "active"
            },
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-config.service": {
                "name": "nfs-config.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-idmap.service": {
                "name": "nfs-idmap.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-lock.service": {
                "name": "nfs-lock.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "nfs-mountd.service": {
                "name": "nfs-mountd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfs-secure-server.service": {
                "name": "nfs-secure-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "nfs-secure.service": {
                "name": "nfs-secure.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "nfs-server.service": {
                "name": "nfs-server.service",
                "source": "systemd",
                "state": "stopped",
                "status": "disabled"
            },
            "nfs-utils.service": {
                "name": "nfs-utils.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "nfs.service": {
                "name": "nfs.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "nfslock.service": {
                "name": "nfslock.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "ntpd.service": {
                "name": "ntpd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "ntpdate.service": {
                "name": "ntpdate.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "ovirt-guest-agent.service": {
                "name": "ovirt-guest-agent.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "plymouth-start.service": {
                "name": "plymouth-start.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "polkit.service": {
                "name": "polkit.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "postfix.service": {
                "name": "postfix.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "quotaon.service": {
                "name": "quotaon.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rc-local.service": {
                "name": "rc-local.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rdisc.service": {
                "name": "rdisc.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rescue.service": {
                "name": "rescue.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rhcd.service": {
                "name": "rhcd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-configure.service": {
                "name": "rhel-configure.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-domainname.service": {
                "name": "rhel-domainname.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-import-state.service": {
                "name": "rhel-import-state.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhel-readonly.service": {
                "name": "rhel-readonly.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "rhnsd": {
                "name": "rhnsd",
                "source": "sysv",
                "state": "running",
                "status": "enabled"
            },
            "rhnsd.service": {
                "name": "rhnsd.service",
                "source": "systemd",
                "state": "running",
                "status": "active"
            },
            "rhsm-facts.service": {
                "name": "rhsm-facts.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rhsm.service": {
                "name": "rhsm.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rhsmcertd.service": {
                "name": "rhsmcertd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rpc-gssd.service": {
                "name": "rpc-gssd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpc-statd.service": {
                "name": "rpc-statd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "rpcbind.service": {
                "name": "rpcbind.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "rpcgssd.service": {
                "name": "rpcgssd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rpcidmapd.service": {
                "name": "rpcidmapd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "rsyncd.service": {
                "name": "rsyncd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "rsyncd@.service": {
                "name": "rsyncd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "rsyslog.service": {
                "name": "rsyslog.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service",
                "source": "systemd",
                "state": "stopped",
                "status": "active"
            },
            "sendmail.service": {
                "name": "sendmail.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "serial-getty@.service": {
                "name": "serial-getty@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service",
                "source": "systemd",
                "state": "running",
                "status": "active"
            },
            "sntp.service": {
                "name": "sntp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "sshd-keygen.service": {
                "name": "sshd-keygen.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "sshd.service": {
                "name": "sshd.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "sshd@.service": {
                "name": "sshd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "syslog.service": {
                "name": "syslog.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-halt.service": {
                "name": "systemd-halt.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-importd.service": {
                "name": "systemd-importd.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-initctl.service": {
                "name": "systemd-initctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-journald.service": {
                "name": "systemd-journald.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-kexec.service": {
                "name": "systemd-kexec.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-localed.service": {
                "name": "systemd-localed.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-logind.service": {
                "name": "systemd-logind.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-machined.service": {
                "name": "systemd-machined.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "disabled"
            },
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "indirect"
            },
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service",
                "source": "systemd",
                "state": "inactive",
                "status": "enabled"
            },
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service",
                "source": "systemd",
                "state": "stopped",
                "status": "enabled"
            },
            "systemd-reboot.service": {
                "name": "systemd-reboot.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-suspend.service": {
                "name": "systemd-suspend.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-sysusers.service": {
                "name": "systemd-sysusers.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "systemd-timedated.service": {
                "name": "systemd-timedated.service",
                "source": "systemd",
                "state": "inactive",
                "status": "static"
            },
            "systemd-timesyncd.service": {
                "name": "systemd-timesyncd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-udevd.service": {
                "name": "systemd-udevd.service",
                "source": "systemd",
                "state": "running",
                "status": "static"
            },
            "systemd-update-done.service": {
                "name": "systemd-update-done.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service",
                "source": "systemd",
                "state": "stopped",
                "status": "static"
            },
            "teamd@.service": {
                "name": "teamd@.service",
                "source": "systemd",
                "state": "unknown",
                "status": "static"
            },
            "tuned.service": {
                "name": "tuned.service",
                "source": "systemd",
                "state": "running",
                "status": "enabled"
            },
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service",
                "source": "systemd",
                "state": "inactive",
                "status": "disabled"
            },
            "ypbind.service": {
                "name": "ypbind.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "yppasswdd.service": {
                "name": "yppasswdd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "ypserv.service": {
                "name": "ypserv.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            },
            "ypxfrd.service": {
                "name": "ypxfrd.service",
                "source": "systemd",
                "state": "stopped",
                "status": "not-found"
            }
        }
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:48:58 +0000 (0:00:01.125)       0:00:14.159 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:48:58 +0000 (0:00:00.083)       0:00:14.242 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:48:58 +0000 (0:00:00.046)       0:00:14.289 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:48:58 +0000 (0:00:00.446)       0:00:14.736 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:48:58 +0000 (0:00:00.034)       0:00:14.770 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:48:58 +0000 (0:00:00.019)       0:00:14.789 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [],
        "changed": false,
        "crypts": [],
        "failed": false,
        "leaves": [],
        "mounts": [],
        "packages": [],
        "pools": [],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:48:58 +0000 (0:00:00.033)       0:00:14.822 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:48:58 +0000 (0:00:00.033)       0:00:14.855 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:48:58 +0000 (0:00:00.033)       0:00:14.888 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:48:59 +0000 (0:00:00.033)       0:00:14.922 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:48:59 +0000 (0:00:00.020)       0:00:14.943 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:48:59 +0000 (0:00:00.031)       0:00:14.974 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:48:59 +0000 (0:00:00.020)       0:00:14.995 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429333.3914723,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:48:59 +0000 (0:00:00.421)       0:00:15.417 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:48:59 +0000 (0:00:00.020)       0:00:15.438 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [Mark tasks to be skipped] ************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:18
Thursday 21 July 2022  18:49:00 +0000 (0:00:00.789)       0:00:16.227 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_skip_checks": [
            "blivet_available",
            "packages_installed",
            "service_facts"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:25
Thursday 21 July 2022  18:49:00 +0000 (0:00:00.031)       0:00:16.259 ********* 
included: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml for /cache/rhel-7.qcow2

TASK [Find unused disks in the system] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml:2
Thursday 21 July 2022  18:49:00 +0000 (0:00:00.033)       0:00:16.293 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "disks": [
        "nvme1n1"
    ]
}

TASK [Set unused_disks if necessary] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml:9
Thursday 21 July 2022  18:49:00 +0000 (0:00:00.423)       0:00:16.716 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "unused_disks": [
            "nvme1n1"
        ]
    },
    "changed": false
}

TASK [Exit playbook when there's not enough unused disks in the system] ********
task path: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml:14
Thursday 21 July 2022  18:49:00 +0000 (0:00:00.034)       0:00:16.750 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Print unused disks] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml:19
Thursday 21 July 2022  18:49:00 +0000 (0:00:00.035)       0:00:16.786 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "unused_disks": [
        "nvme1n1"
    ]
}

TASK [Test creating ext4 filesystem with valid parameter "-Fb 4096"] ***********
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:30
Thursday 21 July 2022  18:49:00 +0000 (0:00:00.059)       0:00:16.845 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:49:00 +0000 (0:00:00.042)       0:00:16.888 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:49:01 +0000 (0:00:00.034)       0:00:16.922 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:49:01 +0000 (0:00:00.433)       0:00:17.356 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:49:01 +0000 (0:00:00.058)       0:00:17.415 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:49:01 +0000 (0:00:00.034)       0:00:17.450 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:49:01 +0000 (0:00:00.032)       0:00:17.482 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:49:01 +0000 (0:00:00.054)       0:00:17.536 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:49:01 +0000 (0:00:00.020)       0:00:17.557 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:49:01 +0000 (0:00:00.025)       0:00:17.582 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_create_options": "-Fb 4096",
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:49:01 +0000 (0:00:00.037)       0:00:17.620 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:49:01 +0000 (0:00:00.034)       0:00:17.655 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:49:01 +0000 (0:00:00.028)       0:00:17.683 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:49:01 +0000 (0:00:00.025)       0:00:17.709 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:49:01 +0000 (0:00:00.025)       0:00:17.734 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:49:01 +0000 (0:00:00.024)       0:00:17.759 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:49:01 +0000 (0:00:00.052)       0:00:17.812 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:49:01 +0000 (0:00:00.019)       0:00:17.831 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/nvme1n1",
            "fs_type": "lvmpv"
        },
        {
            "action": "create device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "create device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd",
        "/dev/mapper/foo-test1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "opts": "defaults",
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs",
        "lvm2"
    ],
    "pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "-Fb 4096",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:49:03 +0000 (0:00:01.655)       0:00:19.487 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:49:03 +0000 (0:00:00.036)       0:00:19.524 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:49:03 +0000 (0:00:00.021)       0:00:19.545 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/nvme1n1",
                "fs_type": "lvmpv"
            },
            {
                "action": "create device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "create device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd",
            "/dev/mapper/foo-test1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "opts": "defaults",
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs",
            "lvm2"
        ],
        "pools": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "-Fb 4096",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:49:03 +0000 (0:00:00.038)       0:00:19.583 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "-Fb 4096",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:49:03 +0000 (0:00:00.041)       0:00:19.625 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:49:03 +0000 (0:00:00.045)       0:00:19.671 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:49:03 +0000 (0:00:00.039)       0:00:19.710 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:49:04 +0000 (0:00:00.748)       0:00:20.458 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'ext4', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "opts": "defaults",
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:49:05 +0000 (0:00:00.497)       0:00:20.955 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:49:05 +0000 (0:00:00.479)       0:00:21.434 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429333.3914723,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:49:05 +0000 (0:00:00.341)       0:00:21.776 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:49:05 +0000 (0:00:00.024)       0:00:21.801 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:45
Thursday 21 July 2022  18:49:07 +0000 (0:00:01.813)       0:00:23.614 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1
Thursday 21 July 2022  18:49:07 +0000 (0:00:00.040)       0:00:23.655 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "-Fb 4096",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6
Thursday 21 July 2022  18:49:07 +0000 (0:00:00.054)       0:00:23.709 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14
Thursday 21 July 2022  18:49:07 +0000 (0:00:00.035)       0:00:23.745 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/mapper/foo-test1": {
            "fstype": "ext4",
            "label": "",
            "name": "/dev/mapper/foo-test1",
            "size": "4G",
            "type": "lvm",
            "uuid": "a69c34d6-bc01-4d83-aad9-b686c548b308"
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "LVM2_member",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": "Z3mwpF-wTUW-OiqS-SSfp-lVpW-yo44-miSJzU"
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-18-48-31-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19
Thursday 21 July 2022  18:49:09 +0000 (0:00:01.423)       0:00:25.168 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003408",
    "end": "2022-07-21 14:49:09.541984",
    "rc": 0,
    "start": "2022-07-21 14:49:09.538576"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0
/dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24
Thursday 21 July 2022  18:49:09 +0000 (0:00:00.418)       0:00:25.587 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003843",
    "end": "2022-07-21 14:49:09.847880",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 14:49:09.844037"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33
Thursday 21 July 2022  18:49:09 +0000 (0:00:00.306)       0:00:25.894 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None})

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:5
Thursday 21 July 2022  18:49:10 +0000 (0:00:00.056)       0:00:25.951 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:18
Thursday 21 July 2022  18:49:10 +0000 (0:00:00.062)       0:00:26.013 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members)
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes)

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:1
Thursday 21 July 2022  18:49:10 +0000 (0:00:00.046)       0:00:26.060 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/nvme1n1"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:6
Thursday 21 July 2022  18:49:10 +0000 (0:00:00.092)       0:00:26.152 ********* 
ok: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/nvme1n1",
    "pv": "/dev/nvme1n1"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:15
Thursday 21 July 2022  18:49:10 +0000 (0:00:00.431)       0:00:26.583 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:19
Thursday 21 July 2022  18:49:10 +0000 (0:00:00.049)       0:00:26.633 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/nvme1n1"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:23
Thursday 21 July 2022  18:49:10 +0000 (0:00:00.049)       0:00:26.682 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:29
Thursday 21 July 2022  18:49:10 +0000 (0:00:00.048)       0:00:26.731 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:33
Thursday 21 July 2022  18:49:10 +0000 (0:00:00.033)       0:00:26.765 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:37
Thursday 21 July 2022  18:49:10 +0000 (0:00:00.045)       0:00:26.811 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:41
Thursday 21 July 2022  18:49:10 +0000 (0:00:00.022)       0:00:26.834 ********* 
ok: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/nvme1n1"
}

MSG:

All assertions passed

TASK [Check MD RAID] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:50
Thursday 21 July 2022  18:49:10 +0000 (0:00:00.037)       0:00:26.871 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml for /cache/rhel-7.qcow2

TASK [get information about RAID] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:6
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.040)       0:00:26.912 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:12
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.026)       0:00:26.938 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:16
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.024)       0:00:26.963 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:20
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.023)       0:00:26.986 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:24
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.022)       0:00:27.008 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:30
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.023)       0:00:27.032 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:36
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.024)       0:00:27.056 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:44
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.023)       0:00:27.080 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:53
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.035)       0:00:27.116 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml:1
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.044)       0:00:27.161 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'})

TASK [Get information about LVM RAID] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.045)       0:00:27.207 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is LVM RAID] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.028)       0:00:27.235 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.029)       0:00:27.265 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:56
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.029)       0:00:27.295 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml for /cache/rhel-7.qcow2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml:1
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.045)       0:00:27.340 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'})

TASK [Get information about thinpool] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.041)       0:00:27.382 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.022)       0:00:27.404 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.025)       0:00:27.430 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.024)       0:00:27.455 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check member encryption] *************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:59
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.024)       0:00:27.479 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:4
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.046)       0:00:27.526 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:8
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.050)       0:00:27.577 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1)  => {
    "_storage_test_pool_member_path": "/dev/nvme1n1",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:15
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.027)       0:00:27.605 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/nvme1n1)

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:1
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.068)       0:00:27.673 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:4
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.050)       0:00:27.723 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:9
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.050)       0:00:27.774 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:15
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.034)       0:00:27.809 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:21
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.039)       0:00:27.849 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:27
Thursday 21 July 2022  18:49:11 +0000 (0:00:00.034)       0:00:27.884 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:22
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.036)       0:00:27.920 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:62
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.038)       0:00:27.959 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml:1
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.050)       0:00:28.009 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'})

TASK [get information about VDO deduplication] *********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.047)       0:00:28.057 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.024)       0:00:28.081 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.023)       0:00:28.105 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.023)       0:00:28.128 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.024)       0:00:28.153 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.022)       0:00:28.176 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.022)       0:00:28.198 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.022)       0:00:28.221 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:65
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.033)       0:00:28.254 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [verify the volumes] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml:3
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.029)       0:00:28.284 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 4096'})

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.038)       0:00:28.323 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.045)       0:00:28.368 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache)

TASK [Get expected mount device based on device type] **************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.078)       0:00:28.447 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.039)       0:00:28.486 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": [
            {
                "block_available": 938726,
                "block_size": 4096,
                "block_total": 999344,
                "block_used": 60618,
                "device": "/dev/mapper/foo-test1",
                "fstype": "ext4",
                "inode_available": 262133,
                "inode_total": 262144,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 3845021696,
                "size_total": 4093313024,
                "uuid": "a69c34d6-bc01-4d83-aad9-b686c548b308"
            }
        ],
        "storage_test_mount_expected_match_count": "1",
        "storage_test_mount_point_matches": [
            {
                "block_available": 938726,
                "block_size": 4096,
                "block_total": 999344,
                "block_used": 60618,
                "device": "/dev/mapper/foo-test1",
                "fstype": "ext4",
                "inode_available": 262133,
                "inode_total": 262144,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 3845021696,
                "size_total": 4093313024,
                "uuid": "a69c34d6-bc01-4d83-aad9-b686c548b308"
            }
        ],
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.062)       0:00:28.548 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the current mount state by mount point] ***************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.059)       0:00:28.608 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the mount fs type] ************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.048)       0:00:28.657 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [command] *****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.050)       0:00:28.707 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.025)       0:00:28.733 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.023)       0:00:28.756 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.021)       0:00:28.778 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": null,
        "storage_test_mount_expected_match_count": null,
        "storage_test_mount_point_matches": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.031)       0:00:28.809 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12
Thursday 21 July 2022  18:49:12 +0000 (0:00:00.056)       0:00:28.866 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19
Thursday 21 July 2022  18:49:13 +0000 (0:00:00.104)       0:00:28.970 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25
Thursday 21 July 2022  18:49:13 +0000 (0:00:00.111)       0:00:29.082 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up variables] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34
Thursday 21 July 2022  18:49:13 +0000 (0:00:00.037)       0:00:29.120 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4
Thursday 21 July 2022  18:49:13 +0000 (0:00:00.035)       0:00:29.155 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10
Thursday 21 July 2022  18:49:13 +0000 (0:00:00.039)       0:00:29.194 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4
Thursday 21 July 2022  18:49:13 +0000 (0:00:00.043)       0:00:29.238 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429343.4074724,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1658429343.4074724,
        "dev": 5,
        "device_type": 64512,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 33267,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1658429343.4074724,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10
Thursday 21 July 2022  18:49:13 +0000 (0:00:00.314)       0:00:29.553 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15
Thursday 21 July 2022  18:49:13 +0000 (0:00:00.038)       0:00:29.592 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [(1/2) Process volume type (set initial value)] ***************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21
Thursday 21 July 2022  18:49:13 +0000 (0:00:00.038)       0:00:29.630 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [(2/2) Process volume type (get RAID value)] ******************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25
Thursday 21 July 2022  18:49:13 +0000 (0:00:00.040)       0:00:29.670 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30
Thursday 21 July 2022  18:49:13 +0000 (0:00:00.024)       0:00:29.695 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3
Thursday 21 July 2022  18:49:13 +0000 (0:00:00.049)       0:00:29.744 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10
Thursday 21 July 2022  18:49:13 +0000 (0:00:00.023)       0:00:29.768 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "changed": true,
    "changes": {
        "installed": [
            "cryptsetup"
        ]
    },
    "rc": 0,
    "results": [
        "Loaded plugins: search-disabled-repos\nResolving Dependencies\n--> Running transaction check\n---> Package cryptsetup.x86_64 0:2.0.3-6.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package             Arch            Version                Repository     Size\n================================================================================\nInstalling:\n cryptsetup          x86_64          2.0.3-6.el7            rhel          154 k\n\nTransaction Summary\n================================================================================\nInstall  1 Package\n\nTotal download size: 154 k\nInstalled size: 354 k\nDownloading packages:\nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n  Installing : cryptsetup-2.0.3-6.el7.x86_64                                1/1 \n  Verifying  : cryptsetup-2.0.3-6.el7.x86_64                                1/1 \n\nInstalled:\n  cryptsetup.x86_64 0:2.0.3-6.el7                                               \n\nComplete!\n"
    ]
}

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15
Thursday 21 July 2022  18:49:15 +0000 (0:00:01.447)       0:00:31.216 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.025)       0:00:31.241 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.024)       0:00:31.265 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.050)       0:00:31.316 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.023)       0:00:31.340 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.024)       0:00:31.365 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.023)       0:00:31.388 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.022)       0:00:31.411 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.022)       0:00:31.434 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.050)       0:00:31.485 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.047)       0:00:31.533 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.035)       0:00:31.568 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.035)       0:00:31.604 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.038)       0:00:31.642 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [get information about RAID] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.033)       0:00:31.676 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.035)       0:00:31.712 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.033)       0:00:31.746 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.036)       0:00:31.783 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.033)       0:00:31.816 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.033)       0:00:31.849 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37
Thursday 21 July 2022  18:49:15 +0000 (0:00:00.034)       0:00:31.884 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the actual size of the volume] *************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3
Thursday 21 July 2022  18:49:16 +0000 (0:00:00.046)       0:00:31.930 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [parse the requested size of the volume] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9
Thursday 21 July 2022  18:49:16 +0000 (0:00:00.458)       0:00:32.389 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15
Thursday 21 July 2022  18:49:16 +0000 (0:00:00.350)       0:00:32.739 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_expected_size": "4294967296"
    },
    "changed": false
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20
Thursday 21 July 2022  18:49:16 +0000 (0:00:00.050)       0:00:32.790 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25
Thursday 21 July 2022  18:49:16 +0000 (0:00:00.036)       0:00:32.826 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28
Thursday 21 July 2022  18:49:16 +0000 (0:00:00.037)       0:00:32.864 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.036)       0:00:32.901 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.038)       0:00:32.940 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.037)       0:00:32.977 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.038)       0:00:33.015 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_actual_size": {
        "bytes": 4294967296,
        "changed": false,
        "failed": false,
        "lvm": "4g",
        "parted": "4GiB",
        "size": "4 GiB"
    }
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.038)       0:00:33.054 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.036)       0:00:33.091 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.048)       0:00:33.139 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.023690",
    "end": "2022-07-21 14:49:17.420153",
    "rc": 0,
    "start": "2022-07-21 14:49:17.396463"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.325)       0:00:33.465 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [check segment type] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.049)       0:00:33.514 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.049)       0:00:33.564 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested cache size] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.036)       0:00:33.600 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.035)       0:00:33.636 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.038)       0:00:33.674 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.038)       0:00:33.712 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.034)       0:00:33.747 ********* 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.020)       0:00:33.768 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Remove the volume group created above] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:47
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.036)       0:00:33.805 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.049)       0:00:33.854 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:49:17 +0000 (0:00:00.038)       0:00:33.892 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:49:18 +0000 (0:00:00.415)       0:00:34.307 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:49:18 +0000 (0:00:00.065)       0:00:34.373 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:49:18 +0000 (0:00:00.066)       0:00:34.439 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:49:18 +0000 (0:00:00.065)       0:00:34.505 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:49:18 +0000 (0:00:00.054)       0:00:34.560 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:49:18 +0000 (0:00:00.019)       0:00:34.580 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:49:18 +0000 (0:00:00.065)       0:00:34.645 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "state": "absent",
            "type": "lvm"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:49:18 +0000 (0:00:00.038)       0:00:34.684 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:49:18 +0000 (0:00:00.037)       0:00:34.722 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:49:18 +0000 (0:00:00.027)       0:00:34.750 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:49:18 +0000 (0:00:00.029)       0:00:34.779 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:49:18 +0000 (0:00:00.028)       0:00:34.808 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:49:18 +0000 (0:00:00.027)       0:00:34.836 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:49:18 +0000 (0:00:00.058)       0:00:34.895 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:49:19 +0000 (0:00:00.022)       0:00:34.918 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/nvme1n1",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "absent",
            "type": "lvm",
            "volumes": []
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:49:20 +0000 (0:00:01.564)       0:00:36.482 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:49:20 +0000 (0:00:00.035)       0:00:36.517 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:49:20 +0000 (0:00:00.022)       0:00:36.540 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/nvme1n1",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:49:20 +0000 (0:00:00.040)       0:00:36.581 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:49:20 +0000 (0:00:00.038)       0:00:36.620 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:49:20 +0000 (0:00:00.036)       0:00:36.656 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:49:21 +0000 (0:00:00.348)       0:00:37.005 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:49:21 +0000 (0:00:00.462)       0:00:37.468 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:49:21 +0000 (0:00:00.039)       0:00:37.507 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:49:22 +0000 (0:00:00.447)       0:00:37.954 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429333.3914723,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:49:22 +0000 (0:00:00.307)       0:00:38.262 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:49:22 +0000 (0:00:00.021)       0:00:38.284 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:57
Thursday 21 July 2022  18:49:23 +0000 (0:00:00.822)       0:00:39.107 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1
Thursday 21 July 2022  18:49:23 +0000 (0:00:00.039)       0:00:39.146 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "absent",
            "type": "lvm",
            "volumes": []
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6
Thursday 21 July 2022  18:49:23 +0000 (0:00:00.076)       0:00:39.222 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14
Thursday 21 July 2022  18:49:23 +0000 (0:00:00.040)       0:00:39.263 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-18-48-31-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19
Thursday 21 July 2022  18:49:24 +0000 (0:00:01.319)       0:00:40.582 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003591",
    "end": "2022-07-21 14:49:24.848411",
    "rc": 0,
    "start": "2022-07-21 14:49:24.844820"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24
Thursday 21 July 2022  18:49:24 +0000 (0:00:00.312)       0:00:40.895 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003541",
    "end": "2022-07-21 14:49:25.164821",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 14:49:25.161280"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.316)       0:00:41.211 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'absent', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None})

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:5
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.057)       0:00:41.269 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:18
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.032)       0:00:41.302 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members)
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes)

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:1
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.052)       0:00:41.354 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "0",
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:6
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.058)       0:00:41.413 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:15
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.023)       0:00:41.436 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": "0"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:19
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.052)       0:00:41.488 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": []
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:23
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.052)       0:00:41.540 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:29
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.049)       0:00:41.590 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:33
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.036)       0:00:41.627 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:37
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.045)       0:00:41.673 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:41
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.025)       0:00:41.698 ********* 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:50
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.022)       0:00:41.721 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml for /cache/rhel-7.qcow2

TASK [get information about RAID] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:6
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.041)       0:00:41.762 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:12
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.025)       0:00:41.788 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:16
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.022)       0:00:41.810 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:20
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.022)       0:00:41.833 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:24
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.027)       0:00:41.860 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:30
Thursday 21 July 2022  18:49:25 +0000 (0:00:00.033)       0:00:41.894 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:36
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.024)       0:00:41.919 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:44
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.025)       0:00:41.945 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:53
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.032)       0:00:41.977 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml:1
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.044)       0:00:42.022 ********* 

TASK [Check Thin Pools] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:56
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.023)       0:00:42.045 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml for /cache/rhel-7.qcow2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml:1
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.076)       0:00:42.122 ********* 

TASK [Check member encryption] *************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:59
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.020)       0:00:42.143 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:4
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.044)       0:00:42.187 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:8
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.049)       0:00:42.237 ********* 

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:15
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.021)       0:00:42.259 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:22
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.022)       0:00:42.281 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:62
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.031)       0:00:42.312 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml:1
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.045)       0:00:42.358 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:65
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.021)       0:00:42.379 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [verify the volumes] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml:3
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.035)       0:00:42.415 ********* 

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.020)       0:00:42.435 ********* 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.020)       0:00:42.456 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Try to create ext4 filesystem with invalid parameter "-Fb 512"] **********
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:61
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.031)       0:00:42.487 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.037)       0:00:42.524 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:49:26 +0000 (0:00:00.033)       0:00:42.558 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:49:27 +0000 (0:00:00.428)       0:00:42.986 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:49:27 +0000 (0:00:00.058)       0:00:43.045 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:49:27 +0000 (0:00:00.035)       0:00:43.081 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:49:27 +0000 (0:00:00.037)       0:00:43.119 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:49:27 +0000 (0:00:00.056)       0:00:43.176 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:49:27 +0000 (0:00:00.019)       0:00:43.195 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:49:27 +0000 (0:00:00.025)       0:00:43.221 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_create_options": "-Fb 512",
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:49:27 +0000 (0:00:00.036)       0:00:43.258 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:49:27 +0000 (0:00:00.032)       0:00:43.291 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:49:27 +0000 (0:00:00.025)       0:00:43.316 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:49:27 +0000 (0:00:00.029)       0:00:43.346 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:49:27 +0000 (0:00:00.032)       0:00:43.379 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:49:27 +0000 (0:00:00.030)       0:00:43.409 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:49:27 +0000 (0:00:00.117)       0:00:43.527 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:49:27 +0000 (0:00:00.024)       0:00:43.551 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [
        "xfsprogs",
        "e2fsprogs",
        "lvm2"
    ],
    "pools": [],
    "volumes": []
}

MSG:

Failed to commit changes to disk: (FSError('format failed: 1',), '/dev/mapper/foo-test1')

TASK [fedora.linux_system_roles.storage : failed message] **********************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:86
Thursday 21 July 2022  18:49:29 +0000 (0:00:01.379)       0:00:44.931 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "changed": false
}

MSG:

{'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': False, 'pools': [{'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': '-Fb 512'}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}], 'volumes': [], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': ['xfsprogs', 'e2fsprogs', 'lvm2'], 'msg': "Failed to commit changes to disk: (FSError('format failed: 1',), '/dev/mapper/foo-test1')", '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:49:29 +0000 (0:00:00.048)       0:00:44.979 ********* 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:81
Thursday 21 July 2022  18:49:29 +0000 (0:00:00.022)       0:00:45.002 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the output when creating ext4 filesystem with invalid parameter "-Fb 512"] ***
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:87
Thursday 21 July 2022  18:49:29 +0000 (0:00:00.034)       0:00:45.036 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Remove the volume group created above] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:92
Thursday 21 July 2022  18:49:29 +0000 (0:00:00.048)       0:00:45.084 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:49:29 +0000 (0:00:00.032)       0:00:45.117 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:49:29 +0000 (0:00:00.034)       0:00:45.152 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:49:29 +0000 (0:00:00.425)       0:00:45.578 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:49:29 +0000 (0:00:00.062)       0:00:45.641 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:49:29 +0000 (0:00:00.036)       0:00:45.678 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:49:29 +0000 (0:00:00.038)       0:00:45.717 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:49:29 +0000 (0:00:00.056)       0:00:45.773 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:49:29 +0000 (0:00:00.020)       0:00:45.793 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:49:29 +0000 (0:00:00.027)       0:00:45.821 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "state": "absent",
            "type": "lvm"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:49:29 +0000 (0:00:00.039)       0:00:45.860 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:49:29 +0000 (0:00:00.034)       0:00:45.894 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:49:30 +0000 (0:00:00.026)       0:00:45.921 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:49:30 +0000 (0:00:00.025)       0:00:45.946 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:49:30 +0000 (0:00:00.029)       0:00:45.976 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:49:30 +0000 (0:00:00.029)       0:00:46.005 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:49:30 +0000 (0:00:00.086)       0:00:46.092 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:49:30 +0000 (0:00:00.022)       0:00:46.115 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/nvme1n1",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "absent",
            "type": "lvm",
            "volumes": []
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:49:31 +0000 (0:00:01.363)       0:00:47.478 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:49:31 +0000 (0:00:00.037)       0:00:47.516 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:49:31 +0000 (0:00:00.021)       0:00:47.537 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/nvme1n1",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:49:31 +0000 (0:00:00.037)       0:00:47.575 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:49:31 +0000 (0:00:00.066)       0:00:47.641 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:49:31 +0000 (0:00:00.067)       0:00:47.709 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:49:31 +0000 (0:00:00.071)       0:00:47.781 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:49:31 +0000 (0:00:00.025)       0:00:47.806 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:49:31 +0000 (0:00:00.033)       0:00:47.839 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:49:31 +0000 (0:00:00.022)       0:00:47.862 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429333.3914723,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:49:32 +0000 (0:00:00.301)       0:00:48.164 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:49:32 +0000 (0:00:00.021)       0:00:48.185 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [Create one LVM logical volume with "4g" under one volume group] **********
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:102
Thursday 21 July 2022  18:49:33 +0000 (0:00:00.816)       0:00:49.001 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:49:33 +0000 (0:00:00.038)       0:00:49.040 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:49:33 +0000 (0:00:00.039)       0:00:49.079 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:49:33 +0000 (0:00:00.414)       0:00:49.493 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:49:33 +0000 (0:00:00.061)       0:00:49.555 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:49:33 +0000 (0:00:00.033)       0:00:49.589 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:49:33 +0000 (0:00:00.038)       0:00:49.628 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:49:33 +0000 (0:00:00.058)       0:00:49.687 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:49:33 +0000 (0:00:00.021)       0:00:49.709 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:49:33 +0000 (0:00:00.029)       0:00:49.739 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "4g"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:49:33 +0000 (0:00:00.041)       0:00:49.780 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:49:33 +0000 (0:00:00.038)       0:00:49.818 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:49:33 +0000 (0:00:00.030)       0:00:49.849 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:49:33 +0000 (0:00:00.031)       0:00:49.881 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:49:34 +0000 (0:00:00.032)       0:00:49.914 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:49:34 +0000 (0:00:00.027)       0:00:49.941 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:49:34 +0000 (0:00:00.090)       0:00:50.032 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:49:34 +0000 (0:00:00.021)       0:00:50.053 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/nvme1n1",
            "fs_type": "lvmpv"
        },
        {
            "action": "create device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "create device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd",
        "/dev/mapper/foo-test1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "opts": "defaults",
            "passno": 0,
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs",
        "lvm2"
    ],
    "pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:49:35 +0000 (0:00:01.578)       0:00:51.632 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:49:35 +0000 (0:00:00.037)       0:00:51.669 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:49:35 +0000 (0:00:00.023)       0:00:51.693 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/nvme1n1",
                "fs_type": "lvmpv"
            },
            {
                "action": "create device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "create device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd",
            "/dev/mapper/foo-test1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "opts": "defaults",
                "passno": 0,
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs",
            "lvm2"
        ],
        "pools": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:49:35 +0000 (0:00:00.043)       0:00:51.736 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "lvm",
                "volumes": [
                    {
                        "_device": "/dev/mapper/foo-test1",
                        "_kernel_device": "/dev/dm-0",
                        "_mount_id": "/dev/mapper/foo-test1",
                        "_raw_device": "/dev/mapper/foo-test1",
                        "_raw_kernel_device": "/dev/dm-0",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": "4g",
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "lvm",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:49:35 +0000 (0:00:00.039)       0:00:51.775 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:49:35 +0000 (0:00:00.035)       0:00:51.810 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:49:35 +0000 (0:00:00.038)       0:00:51.849 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:49:36 +0000 (0:00:00.458)       0:00:52.308 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'ext4', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "opts": "defaults",
        "passno": 0,
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:49:36 +0000 (0:00:00.359)       0:00:52.667 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:49:37 +0000 (0:00:00.451)       0:00:53.118 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429333.3914723,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:49:37 +0000 (0:00:00.306)       0:00:53.425 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:49:37 +0000 (0:00:00.021)       0:00:53.447 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:116
Thursday 21 July 2022  18:49:38 +0000 (0:00:00.839)       0:00:54.286 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1
Thursday 21 July 2022  18:49:38 +0000 (0:00:00.037)       0:00:54.323 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "lvm",
            "volumes": [
                {
                    "_device": "/dev/mapper/foo-test1",
                    "_kernel_device": "/dev/dm-0",
                    "_mount_id": "/dev/mapper/foo-test1",
                    "_raw_device": "/dev/mapper/foo-test1",
                    "_raw_kernel_device": "/dev/dm-0",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": "4g",
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "lvm",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6
Thursday 21 July 2022  18:49:38 +0000 (0:00:00.047)       0:00:54.371 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14
Thursday 21 July 2022  18:49:38 +0000 (0:00:00.035)       0:00:54.406 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/mapper/foo-test1": {
            "fstype": "ext4",
            "label": "",
            "name": "/dev/mapper/foo-test1",
            "size": "4G",
            "type": "lvm",
            "uuid": "a99bc812-ea32-46ee-8400-cdd8d91d3196"
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "LVM2_member",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": "NiVNQk-fzOZ-2IqG-8l37-o8OH-f9cT-rJ9pFZ"
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-18-48-31-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19
Thursday 21 July 2022  18:49:38 +0000 (0:00:00.308)       0:00:54.714 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003676",
    "end": "2022-07-21 14:49:38.961061",
    "rc": 0,
    "start": "2022-07-21 14:49:38.957385"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0
/dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24
Thursday 21 July 2022  18:49:39 +0000 (0:00:00.290)       0:00:55.005 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003264",
    "end": "2022-07-21 14:49:39.253986",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 14:49:39.250722"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33
Thursday 21 July 2022  18:49:39 +0000 (0:00:00.292)       0:00:55.298 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None})

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:5
Thursday 21 July 2022  18:49:39 +0000 (0:00:00.055)       0:00:55.354 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:18
Thursday 21 July 2022  18:49:39 +0000 (0:00:00.032)       0:00:55.387 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members)
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes)

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:1
Thursday 21 July 2022  18:49:39 +0000 (0:00:00.043)       0:00:55.431 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_count": "1",
        "_storage_test_pool_pvs_lvm": [
            "/dev/nvme1n1"
        ]
    },
    "changed": false
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:6
Thursday 21 July 2022  18:49:39 +0000 (0:00:00.116)       0:00:55.547 ********* 
ok: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "device": "/dev/nvme1n1",
    "pv": "/dev/nvme1n1"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:15
Thursday 21 July 2022  18:49:39 +0000 (0:00:00.303)       0:00:55.850 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": "1"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:19
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.049)       0:00:55.900 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_pool_pvs": [
            "/dev/nvme1n1"
        ]
    },
    "changed": false
}

TASK [Verify PV count] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:23
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.054)       0:00:55.954 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:29
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.048)       0:00:56.003 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:33
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.035)       0:00:56.038 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_pv_type": "disk"
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:37
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.048)       0:00:56.087 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:41
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.025)       0:00:56.112 ********* 
ok: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1) => {
    "ansible_loop_var": "pv",
    "changed": false,
    "pv": "/dev/nvme1n1"
}

MSG:

All assertions passed

TASK [Check MD RAID] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:50
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.041)       0:00:56.154 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml for /cache/rhel-7.qcow2

TASK [get information about RAID] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:6
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.040)       0:00:56.194 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:12
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.022)       0:00:56.217 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:16
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.024)       0:00:56.242 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:20
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.022)       0:00:56.265 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:24
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.023)       0:00:56.288 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:30
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.022)       0:00:56.311 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:36
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.024)       0:00:56.335 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:44
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.024)       0:00:56.359 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:53
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.037)       0:00:56.397 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml:1
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.043)       0:00:56.440 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [Get information about LVM RAID] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.042)       0:00:56.483 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is LVM RAID] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.027)       0:00:56.511 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.027)       0:00:56.539 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:56
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.027)       0:00:56.566 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml for /cache/rhel-7.qcow2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml:1
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.043)       0:00:56.610 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [Get information about thinpool] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.039)       0:00:56.649 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in correct thinpool (when thinp name is provided)] ***
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.021)       0:00:56.671 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check that volume is in thinpool (when thinp name is not provided)] ******
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.020)       0:00:56.692 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.020)       0:00:56.713 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check member encryption] *************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:59
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.020)       0:00:56.734 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:4
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.046)       0:00:56.780 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:8
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.084)       0:00:56.864 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1)  => {
    "_storage_test_pool_member_path": "/dev/nvme1n1",
    "ansible_loop_var": "_storage_test_pool_member_path",
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:15
Thursday 21 July 2022  18:49:40 +0000 (0:00:00.031)       0:00:56.896 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/nvme1n1)

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:1
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.090)       0:00:56.987 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": []
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:4
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.048)       0:00:57.035 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:9
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.047)       0:00:57.083 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:15
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.038)       0:00:57.121 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:21
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.036)       0:00:57.158 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:27
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.034)       0:00:57.193 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:22
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.038)       0:00:57.231 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:62
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.031)       0:00:57.263 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml:1
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.050)       0:00:57.313 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [get information about VDO deduplication] *********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.048)       0:00:57.362 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.026)       0:00:57.388 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.025)       0:00:57.414 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.025)       0:00:57.439 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.025)       0:00:57.465 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.025)       0:00:57.490 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.023)       0:00:57.513 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.023)       0:00:57.537 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_vdo_status": null
    },
    "changed": false
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:65
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.033)       0:00:57.571 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [verify the volumes] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml:3
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.035)       0:00:57.606 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.040)       0:00:57.647 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.048)       0:00:57.696 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache)

TASK [Get expected mount device based on device type] **************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.075)       0:00:57.771 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/mapper/foo-test1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.039)       0:00:57.810 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": [
            {
                "block_available": 938726,
                "block_size": 4096,
                "block_total": 999344,
                "block_used": 60618,
                "device": "/dev/mapper/foo-test1",
                "fstype": "ext4",
                "inode_available": 262133,
                "inode_total": 262144,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 3845021696,
                "size_total": 4093313024,
                "uuid": "a99bc812-ea32-46ee-8400-cdd8d91d3196"
            }
        ],
        "storage_test_mount_expected_match_count": "1",
        "storage_test_mount_point_matches": [
            {
                "block_available": 938726,
                "block_size": 4096,
                "block_total": 999344,
                "block_used": 60618,
                "device": "/dev/mapper/foo-test1",
                "fstype": "ext4",
                "inode_available": 262133,
                "inode_total": 262144,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 3845021696,
                "size_total": 4093313024,
                "uuid": "a99bc812-ea32-46ee-8400-cdd8d91d3196"
            }
        ],
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20
Thursday 21 July 2022  18:49:41 +0000 (0:00:00.058)       0:00:57.868 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the current mount state by mount point] ***************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29
Thursday 21 July 2022  18:49:42 +0000 (0:00:00.052)       0:00:57.921 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the mount fs type] ************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37
Thursday 21 July 2022  18:49:42 +0000 (0:00:00.049)       0:00:57.970 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [command] *****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46
Thursday 21 July 2022  18:49:42 +0000 (0:00:00.052)       0:00:58.023 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50
Thursday 21 July 2022  18:49:42 +0000 (0:00:00.024)       0:00:58.048 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55
Thursday 21 July 2022  18:49:42 +0000 (0:00:00.023)       0:00:58.071 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65
Thursday 21 July 2022  18:49:42 +0000 (0:00:00.022)       0:00:58.093 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": null,
        "storage_test_mount_expected_match_count": null,
        "storage_test_mount_point_matches": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2
Thursday 21 July 2022  18:49:42 +0000 (0:00:00.034)       0:00:58.128 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "/dev/mapper/foo-test1 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12
Thursday 21 July 2022  18:49:42 +0000 (0:00:00.061)       0:00:58.189 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19
Thursday 21 July 2022  18:49:42 +0000 (0:00:00.084)       0:00:58.274 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25
Thursday 21 July 2022  18:49:42 +0000 (0:00:00.117)       0:00:58.391 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up variables] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34
Thursday 21 July 2022  18:49:42 +0000 (0:00:00.037)       0:00:58.428 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4
Thursday 21 July 2022  18:49:42 +0000 (0:00:00.036)       0:00:58.464 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10
Thursday 21 July 2022  18:49:42 +0000 (0:00:00.041)       0:00:58.505 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4
Thursday 21 July 2022  18:49:42 +0000 (0:00:00.042)       0:00:58.548 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429375.5084722,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1658429375.5084722,
        "dev": 5,
        "device_type": 64512,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 41418,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/symlink",
        "mode": "0660",
        "mtime": 1658429375.5084722,
        "nlink": 1,
        "path": "/dev/mapper/foo-test1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10
Thursday 21 July 2022  18:49:42 +0000 (0:00:00.311)       0:00:58.860 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15
Thursday 21 July 2022  18:49:42 +0000 (0:00:00.037)       0:00:58.897 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [(1/2) Process volume type (set initial value)] ***************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21
Thursday 21 July 2022  18:49:43 +0000 (0:00:00.040)       0:00:58.938 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "st_volume_type": "lvm"
    },
    "changed": false
}

TASK [(2/2) Process volume type (get RAID value)] ******************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25
Thursday 21 July 2022  18:49:43 +0000 (0:00:00.036)       0:00:58.974 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30
Thursday 21 July 2022  18:49:43 +0000 (0:00:00.023)       0:00:58.998 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3
Thursday 21 July 2022  18:49:43 +0000 (0:00:00.040)       0:00:59.038 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10
Thursday 21 July 2022  18:49:43 +0000 (0:00:00.023)       0:00:59.062 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "rc": 0,
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15
Thursday 21 July 2022  18:49:43 +0000 (0:00:00.527)       0:00:59.590 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21
Thursday 21 July 2022  18:49:43 +0000 (0:00:00.025)       0:00:59.615 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27
Thursday 21 July 2022  18:49:43 +0000 (0:00:00.024)       0:00:59.640 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33
Thursday 21 July 2022  18:49:43 +0000 (0:00:00.052)       0:00:59.692 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39
Thursday 21 July 2022  18:49:43 +0000 (0:00:00.025)       0:00:59.718 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44
Thursday 21 July 2022  18:49:43 +0000 (0:00:00.024)       0:00:59.742 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50
Thursday 21 July 2022  18:49:43 +0000 (0:00:00.026)       0:00:59.768 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56
Thursday 21 July 2022  18:49:43 +0000 (0:00:00.025)       0:00:59.794 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62
Thursday 21 July 2022  18:49:43 +0000 (0:00:00.024)       0:00:59.819 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67
Thursday 21 July 2022  18:49:43 +0000 (0:00:00.056)       0:00:59.875 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72
Thursday 21 July 2022  18:49:44 +0000 (0:00:00.052)       0:00:59.928 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78
Thursday 21 July 2022  18:49:44 +0000 (0:00:00.039)       0:00:59.967 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84
Thursday 21 July 2022  18:49:44 +0000 (0:00:00.039)       0:01:00.007 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90
Thursday 21 July 2022  18:49:44 +0000 (0:00:00.036)       0:01:00.043 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [get information about RAID] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7
Thursday 21 July 2022  18:49:44 +0000 (0:00:00.036)       0:01:00.079 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13
Thursday 21 July 2022  18:49:44 +0000 (0:00:00.037)       0:01:00.117 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17
Thursday 21 July 2022  18:49:44 +0000 (0:00:00.037)       0:01:00.155 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21
Thursday 21 July 2022  18:49:44 +0000 (0:00:00.036)       0:01:00.191 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25
Thursday 21 July 2022  18:49:44 +0000 (0:00:00.039)       0:01:00.231 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31
Thursday 21 July 2022  18:49:44 +0000 (0:00:00.036)       0:01:00.267 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37
Thursday 21 July 2022  18:49:44 +0000 (0:00:00.038)       0:01:00.306 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the actual size of the volume] *************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3
Thursday 21 July 2022  18:49:44 +0000 (0:00:00.037)       0:01:00.343 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [parse the requested size of the volume] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9
Thursday 21 July 2022  18:49:44 +0000 (0:00:00.320)       0:01:00.664 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "bytes": 4294967296,
    "changed": false,
    "lvm": "4g",
    "parted": "4GiB",
    "size": "4 GiB"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15
Thursday 21 July 2022  18:49:45 +0000 (0:00:00.327)       0:01:00.991 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_expected_size": "4294967296"
    },
    "changed": false
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20
Thursday 21 July 2022  18:49:45 +0000 (0:00:00.098)       0:01:01.090 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25
Thursday 21 July 2022  18:49:45 +0000 (0:00:00.037)       0:01:01.127 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28
Thursday 21 July 2022  18:49:45 +0000 (0:00:00.038)       0:01:01.166 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31
Thursday 21 July 2022  18:49:45 +0000 (0:00:00.037)       0:01:01.203 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36
Thursday 21 July 2022  18:49:45 +0000 (0:00:00.036)       0:01:01.240 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39
Thursday 21 July 2022  18:49:45 +0000 (0:00:00.046)       0:01:01.287 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44
Thursday 21 July 2022  18:49:45 +0000 (0:00:00.076)       0:01:01.364 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_actual_size": {
        "bytes": 4294967296,
        "changed": false,
        "failed": false,
        "lvm": "4g",
        "parted": "4GiB",
        "size": "4 GiB"
    }
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47
Thursday 21 July 2022  18:49:45 +0000 (0:00:00.040)       0:01:01.404 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50
Thursday 21 July 2022  18:49:45 +0000 (0:00:00.037)       0:01:01.442 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get information about the LV] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6
Thursday 21 July 2022  18:49:45 +0000 (0:00:00.056)       0:01:01.499 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "lvs",
        "--noheadings",
        "--nameprefixes",
        "--units=b",
        "--nosuffix",
        "--unquoted",
        "-o",
        "name,attr,cache_total_blocks,chunk_size,segtype",
        "foo/test1"
    ],
    "delta": "0:00:00.023195",
    "end": "2022-07-21 14:49:45.774838",
    "rc": 0,
    "start": "2022-07-21 14:49:45.751643"
}

STDOUT:

  LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14
Thursday 21 July 2022  18:49:45 +0000 (0:00:00.324)       0:01:01.823 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_lv_segtype": [
            "linear"
        ]
    },
    "changed": false
}

TASK [check segment type] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17
Thursday 21 July 2022  18:49:45 +0000 (0:00:00.051)       0:01:01.875 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22
Thursday 21 July 2022  18:49:46 +0000 (0:00:00.052)       0:01:01.928 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested cache size] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26
Thursday 21 July 2022  18:49:46 +0000 (0:00:00.041)       0:01:01.969 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32
Thursday 21 July 2022  18:49:46 +0000 (0:00:00.036)       0:01:02.005 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36
Thursday 21 July 2022  18:49:46 +0000 (0:00:00.035)       0:01:02.041 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16
Thursday 21 July 2022  18:49:46 +0000 (0:00:00.036)       0:01:02.077 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43
Thursday 21 July 2022  18:49:46 +0000 (0:00:00.036)       0:01:02.113 ********* 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53
Thursday 21 July 2022  18:49:46 +0000 (0:00:00.021)       0:01:02.135 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Resizing with one large value which large than disk's size] **************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:120
Thursday 21 July 2022  18:49:46 +0000 (0:00:00.037)       0:01:02.172 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:49:46 +0000 (0:00:00.039)       0:01:02.211 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:49:46 +0000 (0:00:00.036)       0:01:02.248 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:49:46 +0000 (0:00:00.411)       0:01:02.660 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:49:46 +0000 (0:00:00.062)       0:01:02.722 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:49:46 +0000 (0:00:00.035)       0:01:02.757 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:49:46 +0000 (0:00:00.041)       0:01:02.799 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:49:46 +0000 (0:00:00.063)       0:01:02.862 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:49:46 +0000 (0:00:00.021)       0:01:02.883 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:49:47 +0000 (0:00:00.035)       0:01:02.919 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "type": "lvm",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "size": "12884901888.0"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:49:47 +0000 (0:00:00.058)       0:01:02.978 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:49:47 +0000 (0:00:00.041)       0:01:03.019 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:49:47 +0000 (0:00:00.032)       0:01:03.051 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:49:47 +0000 (0:00:00.029)       0:01:03.080 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:49:47 +0000 (0:00:00.029)       0:01:03.110 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:49:47 +0000 (0:00:00.029)       0:01:03.140 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:49:47 +0000 (0:00:00.104)       0:01:03.244 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:49:47 +0000 (0:00:00.023)       0:01:03.268 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

volume 'test1' cannot be resized to '12 GiB'

TASK [fedora.linux_system_roles.storage : failed message] **********************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:86
Thursday 21 July 2022  18:49:48 +0000 (0:00:01.212)       0:01:04.480 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "changed": false
}

MSG:

{'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': False, 'pools': [{'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'lvm', 'size': '12884901888.0', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}], 'volumes': [], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': [], 'msg': "volume 'test1' cannot be resized to '12 GiB'", '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:49:48 +0000 (0:00:00.123)       0:01:04.603 ********* 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:139
Thursday 21 July 2022  18:49:48 +0000 (0:00:00.022)       0:01:04.626 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the output when resizing with large size] *************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:145
Thursday 21 July 2022  18:49:48 +0000 (0:00:00.036)       0:01:04.662 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Remove the volume group created above] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:150
Thursday 21 July 2022  18:49:48 +0000 (0:00:00.049)       0:01:04.711 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:49:48 +0000 (0:00:00.037)       0:01:04.748 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:49:48 +0000 (0:00:00.034)       0:01:04.782 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:49:49 +0000 (0:00:00.413)       0:01:05.196 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:49:49 +0000 (0:00:00.063)       0:01:05.259 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:49:49 +0000 (0:00:00.037)       0:01:05.296 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:49:49 +0000 (0:00:00.036)       0:01:05.333 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:49:49 +0000 (0:00:00.056)       0:01:05.389 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:49:49 +0000 (0:00:00.021)       0:01:05.411 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:49:49 +0000 (0:00:00.026)       0:01:05.438 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "state": "absent",
            "type": "lvm"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:49:49 +0000 (0:00:00.039)       0:01:05.478 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:49:49 +0000 (0:00:00.040)       0:01:05.518 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:49:49 +0000 (0:00:00.031)       0:01:05.549 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:49:49 +0000 (0:00:00.029)       0:01:05.579 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:49:49 +0000 (0:00:00.030)       0:01:05.610 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:49:49 +0000 (0:00:00.029)       0:01:05.639 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:49:49 +0000 (0:00:00.059)       0:01:05.698 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:49:49 +0000 (0:00:00.021)       0:01:05.720 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/mapper/foo-test1",
            "fs_type": "ext4"
        },
        {
            "action": "destroy device",
            "device": "/dev/mapper/foo-test1",
            "fs_type": null
        },
        {
            "action": "destroy device",
            "device": "/dev/foo",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/nvme1n1",
            "fs_type": "lvmpv"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "/dev/mapper/foo-test1",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "absent",
            "type": "lvm",
            "volumes": []
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:49:51 +0000 (0:00:01.515)       0:01:07.236 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:49:51 +0000 (0:00:00.037)       0:01:07.273 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:49:51 +0000 (0:00:00.021)       0:01:07.295 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/mapper/foo-test1",
                "fs_type": "ext4"
            },
            {
                "action": "destroy device",
                "device": "/dev/mapper/foo-test1",
                "fs_type": null
            },
            {
                "action": "destroy device",
                "device": "/dev/foo",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/nvme1n1",
                "fs_type": "lvmpv"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "/dev/mapper/foo-test1",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:49:51 +0000 (0:00:00.041)       0:01:07.336 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "absent",
                "type": "lvm",
                "volumes": []
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:49:51 +0000 (0:00:00.037)       0:01:07.374 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:49:51 +0000 (0:00:00.039)       0:01:07.413 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "/dev/mapper/foo-test1",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "/dev/mapper/foo-test1"
}

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:49:51 +0000 (0:00:00.371)       0:01:07.784 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:49:52 +0000 (0:00:00.469)       0:01:08.254 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:49:52 +0000 (0:00:00.039)       0:01:08.294 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:49:52 +0000 (0:00:00.475)       0:01:08.769 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429333.3914723,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:49:53 +0000 (0:00:00.318)       0:01:09.088 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:49:53 +0000 (0:00:00.030)       0:01:09.118 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [Create one partition on one disk] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:160
Thursday 21 July 2022  18:49:54 +0000 (0:00:00.811)       0:01:09.929 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:49:54 +0000 (0:00:00.042)       0:01:09.972 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:49:54 +0000 (0:00:00.034)       0:01:10.007 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:49:54 +0000 (0:00:00.422)       0:01:10.429 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:49:54 +0000 (0:00:00.064)       0:01:10.494 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:49:54 +0000 (0:00:00.039)       0:01:10.533 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:49:54 +0000 (0:00:00.034)       0:01:10.568 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:49:54 +0000 (0:00:00.054)       0:01:10.622 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:49:54 +0000 (0:00:00.021)       0:01:10.644 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:49:54 +0000 (0:00:00.027)       0:01:10.671 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "type": "partition",
            "volumes": [
                {
                    "fs_type": "ext4",
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "type": "partition"
                }
            ]
        }
    ]
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:49:54 +0000 (0:00:00.036)       0:01:10.708 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined"
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:49:54 +0000 (0:00:00.034)       0:01:10.742 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:49:54 +0000 (0:00:00.031)       0:01:10.774 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:49:54 +0000 (0:00:00.029)       0:01:10.804 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:49:54 +0000 (0:00:00.027)       0:01:10.832 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:49:54 +0000 (0:00:00.026)       0:01:10.858 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:49:55 +0000 (0:00:00.057)       0:01:10.915 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:49:55 +0000 (0:00:00.020)       0:01:10.936 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "create format",
            "device": "/dev/nvme1n1",
            "fs_type": "disklabel"
        },
        {
            "action": "create device",
            "device": "/dev/nvme1n1p1",
            "fs_type": null
        },
        {
            "action": "create format",
            "device": "/dev/nvme1n1p1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd",
        "/dev/nvme1n1p1"
    ],
    "mounts": [
        {
            "dump": 0,
            "fstype": "ext4",
            "opts": "defaults",
            "passno": 0,
            "path": "/opt/test1",
            "src": "UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs"
    ],
    "pools": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "partition",
            "volumes": [
                {
                    "_device": "/dev/nvme1n1p1",
                    "_kernel_device": "/dev/nvme1n1p1",
                    "_mount_id": "UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6",
                    "_raw_device": "/dev/nvme1n1p1",
                    "_raw_kernel_device": "/dev/nvme1n1p1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": 0,
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "partition",
                    "vdo_pool_size": null
                }
            ]
        }
    ],
    "volumes": []
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:49:56 +0000 (0:00:01.542)       0:01:12.478 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:49:56 +0000 (0:00:00.038)       0:01:12.517 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:49:56 +0000 (0:00:00.021)       0:01:12.539 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "create format",
                "device": "/dev/nvme1n1",
                "fs_type": "disklabel"
            },
            {
                "action": "create device",
                "device": "/dev/nvme1n1p1",
                "fs_type": null
            },
            {
                "action": "create format",
                "device": "/dev/nvme1n1p1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd",
            "/dev/nvme1n1p1"
        ],
        "mounts": [
            {
                "dump": 0,
                "fstype": "ext4",
                "opts": "defaults",
                "passno": 0,
                "path": "/opt/test1",
                "src": "UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs"
        ],
        "pools": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "partition",
                "volumes": [
                    {
                        "_device": "/dev/nvme1n1p1",
                        "_kernel_device": "/dev/nvme1n1p1",
                        "_mount_id": "UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6",
                        "_raw_device": "/dev/nvme1n1p1",
                        "_raw_kernel_device": "/dev/nvme1n1p1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": 0,
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "partition",
                        "vdo_pool_size": null
                    }
                ]
            }
        ],
        "volumes": []
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:49:56 +0000 (0:00:00.037)       0:01:12.577 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": [
            {
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "state": "present",
                "type": "partition",
                "volumes": [
                    {
                        "_device": "/dev/nvme1n1p1",
                        "_kernel_device": "/dev/nvme1n1p1",
                        "_mount_id": "UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6",
                        "_raw_device": "/dev/nvme1n1p1",
                        "_raw_kernel_device": "/dev/nvme1n1p1",
                        "cache_devices": [],
                        "cache_mode": null,
                        "cache_size": 0,
                        "cached": false,
                        "compression": null,
                        "deduplication": null,
                        "disks": [],
                        "encryption": false,
                        "encryption_cipher": null,
                        "encryption_key": null,
                        "encryption_key_size": null,
                        "encryption_luks_version": null,
                        "encryption_password": null,
                        "fs_create_options": "",
                        "fs_label": "",
                        "fs_overwrite_existing": true,
                        "fs_type": "ext4",
                        "mount_check": 0,
                        "mount_device_identifier": "uuid",
                        "mount_options": "defaults",
                        "mount_passno": 0,
                        "mount_point": "/opt/test1",
                        "name": "test1",
                        "raid_chunk_size": null,
                        "raid_device_count": null,
                        "raid_disks": [],
                        "raid_level": null,
                        "raid_metadata_version": null,
                        "raid_spare_count": null,
                        "size": 0,
                        "state": "present",
                        "thin": false,
                        "thin_pool_name": null,
                        "thin_pool_size": null,
                        "type": "partition",
                        "vdo_pool_size": null
                    }
                ]
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:49:56 +0000 (0:00:00.071)       0:01:12.649 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:49:56 +0000 (0:00:00.070)       0:01:12.719 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:49:56 +0000 (0:00:00.040)       0:01:12.760 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:49:57 +0000 (0:00:00.502)       0:01:13.262 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': 'UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6', 'dump': 0, 'passno': 0, 'fstype': 'ext4', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "opts": "defaults",
        "passno": 0,
        "path": "/opt/test1",
        "src": "UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6"
}

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:49:57 +0000 (0:00:00.338)       0:01:13.601 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:49:58 +0000 (0:00:00.449)       0:01:14.050 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429333.3914723,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:49:58 +0000 (0:00:00.316)       0:01:14.367 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:49:58 +0000 (0:00:00.023)       0:01:14.390 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:174
Thursday 21 July 2022  18:49:59 +0000 (0:00:00.821)       0:01:15.211 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1
Thursday 21 July 2022  18:49:59 +0000 (0:00:00.037)       0:01:15.249 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_pools_list": [
        {
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "state": "present",
            "type": "partition",
            "volumes": [
                {
                    "_device": "/dev/nvme1n1p1",
                    "_kernel_device": "/dev/nvme1n1p1",
                    "_mount_id": "UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6",
                    "_raw_device": "/dev/nvme1n1p1",
                    "_raw_kernel_device": "/dev/nvme1n1p1",
                    "cache_devices": [],
                    "cache_mode": null,
                    "cache_size": 0,
                    "cached": false,
                    "compression": null,
                    "deduplication": null,
                    "disks": [],
                    "encryption": false,
                    "encryption_cipher": null,
                    "encryption_key": null,
                    "encryption_key_size": null,
                    "encryption_luks_version": null,
                    "encryption_password": null,
                    "fs_create_options": "",
                    "fs_label": "",
                    "fs_overwrite_existing": true,
                    "fs_type": "ext4",
                    "mount_check": 0,
                    "mount_device_identifier": "uuid",
                    "mount_options": "defaults",
                    "mount_passno": 0,
                    "mount_point": "/opt/test1",
                    "name": "test1",
                    "raid_chunk_size": null,
                    "raid_device_count": null,
                    "raid_disks": [],
                    "raid_level": null,
                    "raid_metadata_version": null,
                    "raid_spare_count": null,
                    "size": 0,
                    "state": "present",
                    "thin": false,
                    "thin_pool_name": null,
                    "thin_pool_size": null,
                    "type": "partition",
                    "vdo_pool_size": null
                }
            ]
        }
    ]
}

TASK [Print out volume information] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6
Thursday 21 July 2022  18:49:59 +0000 (0:00:00.050)       0:01:15.299 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14
Thursday 21 July 2022  18:49:59 +0000 (0:00:00.034)       0:01:15.334 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1p1": {
            "fstype": "ext4",
            "label": "",
            "name": "/dev/nvme1n1p1",
            "size": "10G",
            "type": "partition",
            "uuid": "9ce48a48-ec04-43d1-aaf5-dd298f4a75f6"
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-18-48-31-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19
Thursday 21 July 2022  18:49:59 +0000 (0:00:00.312)       0:01:15.647 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003406",
    "end": "2022-07-21 14:49:59.903984",
    "rc": 0,
    "start": "2022-07-21 14:49:59.900578"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0
UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6 /opt/test1 ext4 defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.302)       0:01:15.950 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.003585",
    "end": "2022-07-21 14:50:00.211059",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 14:50:00.207474"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.307)       0:01:16.258 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/nvme1n1p1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/nvme1n1p1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/nvme1n1p1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/nvme1n1p1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'partition', 'encryption_cipher': None, 'raid_spare_count': None})

TASK [Set _storage_pool_tests] *************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:5
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.062)       0:01:16.320 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pool_tests": [
            "members",
            "volumes"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:18
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.085)       0:01:16.406 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members)
included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes)

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:1
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.046)       0:01:16.453 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get the canonical device path for each member device] ********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:6
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.023)       0:01:16.476 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:15
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.020)       0:01:16.496 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:19
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.025)       0:01:16.522 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify PV count] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:23
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.024)       0:01:16.546 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:29
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.024)       0:01:16.571 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:33
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.023)       0:01:16.594 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:37
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.065)       0:01:16.660 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check the type of each PV] ***********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:41
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.024)       0:01:16.685 ********* 

TASK [Check MD RAID] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:50
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.022)       0:01:16.707 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml for /cache/rhel-7.qcow2

TASK [get information about RAID] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:6
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.041)       0:01:16.749 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:12
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.024)       0:01:16.773 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:16
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.024)       0:01:16.797 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:20
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.024)       0:01:16.822 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:24
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.024)       0:01:16.846 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:30
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.023)       0:01:16.870 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:36
Thursday 21 July 2022  18:50:00 +0000 (0:00:00.022)       0:01:16.893 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:44
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.024)       0:01:16.917 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_md_active_devices_re": null,
        "storage_test_md_metadata_version_re": null,
        "storage_test_md_spare_devices_re": null
    },
    "changed": false
}

TASK [Check LVM RAID] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:53
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.033)       0:01:16.951 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2

TASK [Validate pool member LVM RAID settings] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml:1
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.044)       0:01:16.996 ********* 
skipping: [/cache/rhel-7.qcow2] => (item={'_raw_device': '/dev/nvme1n1p1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/nvme1n1p1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/nvme1n1p1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/nvme1n1p1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})  => {
    "ansible_loop_var": "storage_test_lvmraid_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_lvmraid_volume": {
        "_device": "/dev/nvme1n1p1",
        "_kernel_device": "/dev/nvme1n1p1",
        "_mount_id": "UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6",
        "_raw_device": "/dev/nvme1n1p1",
        "_raw_kernel_device": "/dev/nvme1n1p1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [],
        "encryption": false,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": null,
        "encryption_luks_version": null,
        "encryption_password": null,
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "ext4",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "size": 0,
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check Thin Pools] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:56
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.033)       0:01:17.030 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml for /cache/rhel-7.qcow2

TASK [Validate pool member thinpool settings] **********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml:1
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.044)       0:01:17.074 ********* 
skipping: [/cache/rhel-7.qcow2] => (item={'_raw_device': '/dev/nvme1n1p1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/nvme1n1p1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/nvme1n1p1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/nvme1n1p1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})  => {
    "ansible_loop_var": "storage_test_thin_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_thin_volume": {
        "_device": "/dev/nvme1n1p1",
        "_kernel_device": "/dev/nvme1n1p1",
        "_mount_id": "UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6",
        "_raw_device": "/dev/nvme1n1p1",
        "_raw_kernel_device": "/dev/nvme1n1p1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [],
        "encryption": false,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": null,
        "encryption_luks_version": null,
        "encryption_password": null,
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "ext4",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "size": 0,
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [Check member encryption] *************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:59
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.030)       0:01:17.105 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:4
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.048)       0:01:17.153 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Validate pool member LUKS settings] **************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:8
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.048)       0:01:17.201 ********* 

TASK [Validate pool member crypttab entries] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:15
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.022)       0:01:17.223 ********* 

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:22
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.022)       0:01:17.246 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_crypttab_key_file": null
    },
    "changed": false
}

TASK [Check VDO] ***************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:62
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.034)       0:01:17.280 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2

TASK [Validate pool member VDO settings] ***************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml:1
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.046)       0:01:17.327 ********* 
skipping: [/cache/rhel-7.qcow2] => (item={'_raw_device': '/dev/nvme1n1p1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/nvme1n1p1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/nvme1n1p1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/nvme1n1p1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})  => {
    "ansible_loop_var": "storage_test_vdo_volume",
    "changed": false,
    "skip_reason": "Conditional result was False",
    "storage_test_vdo_volume": {
        "_device": "/dev/nvme1n1p1",
        "_kernel_device": "/dev/nvme1n1p1",
        "_mount_id": "UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6",
        "_raw_device": "/dev/nvme1n1p1",
        "_raw_kernel_device": "/dev/nvme1n1p1",
        "cache_devices": [],
        "cache_mode": null,
        "cache_size": 0,
        "cached": false,
        "compression": null,
        "deduplication": null,
        "disks": [],
        "encryption": false,
        "encryption_cipher": null,
        "encryption_key": null,
        "encryption_key_size": null,
        "encryption_luks_version": null,
        "encryption_password": null,
        "fs_create_options": "",
        "fs_label": "",
        "fs_overwrite_existing": true,
        "fs_type": "ext4",
        "mount_check": 0,
        "mount_device_identifier": "uuid",
        "mount_options": "defaults",
        "mount_passno": 0,
        "mount_point": "/opt/test1",
        "name": "test1",
        "raid_chunk_size": null,
        "raid_device_count": null,
        "raid_disks": [],
        "raid_level": null,
        "raid_metadata_version": null,
        "raid_spare_count": null,
        "size": 0,
        "state": "present",
        "thin": false,
        "thin_pool_name": null,
        "thin_pool_size": null,
        "type": "partition",
        "vdo_pool_size": null
    }
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:65
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.029)       0:01:17.356 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "__pvs_lvm_len": null,
        "_storage_test_expected_pv_count": null,
        "_storage_test_expected_pv_type": null,
        "_storage_test_pool_pvs": [],
        "_storage_test_pool_pvs_lvm": []
    },
    "changed": false
}

TASK [verify the volumes] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml:3
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.035)       0:01:17.392 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/nvme1n1p1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'ext4', 'mount_options': 'defaults', 'type': 'partition', '_device': '/dev/nvme1n1p1', 'size': 0, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/nvme1n1p1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/nvme1n1p1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''})

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.042)       0:01:17.434 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.049)       0:01:17.483 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache)

TASK [Get expected mount device based on device type] **************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.078)       0:01:17.562 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/nvme1n1p1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.042)       0:01:17.604 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": [
            {
                "block_available": 2402892,
                "block_size": 4096,
                "block_total": 2547269,
                "block_used": 144377,
                "device": "/dev/nvme1n1p1",
                "fstype": "ext4",
                "inode_available": 655349,
                "inode_total": 655360,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 9842245632,
                "size_total": 10433613824,
                "uuid": "9ce48a48-ec04-43d1-aaf5-dd298f4a75f6"
            }
        ],
        "storage_test_mount_expected_match_count": "1",
        "storage_test_mount_point_matches": [
            {
                "block_available": 2402892,
                "block_size": 4096,
                "block_total": 2547269,
                "block_used": 144377,
                "device": "/dev/nvme1n1p1",
                "fstype": "ext4",
                "inode_available": 655349,
                "inode_total": 655360,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,relatime,data=ordered",
                "size_available": 9842245632,
                "size_total": 10433613824,
                "uuid": "9ce48a48-ec04-43d1-aaf5-dd298f4a75f6"
            }
        ],
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.065)       0:01:17.670 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the current mount state by mount point] ***************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.052)       0:01:17.723 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the mount fs type] ************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37
Thursday 21 July 2022  18:50:01 +0000 (0:00:00.086)       0:01:17.809 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [command] *****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.134)       0:01:17.944 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.026)       0:01:17.970 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.024)       0:01:17.995 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.023)       0:01:18.019 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": null,
        "storage_test_mount_expected_match_count": null,
        "storage_test_mount_point_matches": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.038)       0:01:18.057 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6 "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.066)       0:01:18.123 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.050)       0:01:18.174 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.063)       0:01:18.237 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up variables] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.039)       0:01:18.276 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.035)       0:01:18.312 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.039)       0:01:18.351 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.043)       0:01:18.395 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429396.3904722,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1658429396.3904722,
        "dev": 5,
        "device_type": 66307,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 50532,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/blockdevice",
        "mode": "0660",
        "mtime": 1658429396.3904722,
        "nlink": 1,
        "path": "/dev/nvme1n1p1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.322)       0:01:18.718 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.040)       0:01:18.758 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [(1/2) Process volume type (set initial value)] ***************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.040)       0:01:18.798 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "st_volume_type": "partition"
    },
    "changed": false
}

TASK [(2/2) Process volume type (get RAID value)] ******************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.042)       0:01:18.841 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30
Thursday 21 July 2022  18:50:02 +0000 (0:00:00.025)       0:01:18.867 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3
Thursday 21 July 2022  18:50:03 +0000 (0:00:00.040)       0:01:18.907 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10
Thursday 21 July 2022  18:50:03 +0000 (0:00:00.025)       0:01:18.932 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "rc": 0,
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15
Thursday 21 July 2022  18:50:03 +0000 (0:00:00.543)       0:01:19.476 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21
Thursday 21 July 2022  18:50:03 +0000 (0:00:00.026)       0:01:19.502 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27
Thursday 21 July 2022  18:50:03 +0000 (0:00:00.023)       0:01:19.525 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33
Thursday 21 July 2022  18:50:03 +0000 (0:00:00.050)       0:01:19.576 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39
Thursday 21 July 2022  18:50:03 +0000 (0:00:00.028)       0:01:19.604 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44
Thursday 21 July 2022  18:50:03 +0000 (0:00:00.024)       0:01:19.628 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50
Thursday 21 July 2022  18:50:03 +0000 (0:00:00.024)       0:01:19.653 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56
Thursday 21 July 2022  18:50:03 +0000 (0:00:00.023)       0:01:19.676 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62
Thursday 21 July 2022  18:50:03 +0000 (0:00:00.022)       0:01:19.699 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67
Thursday 21 July 2022  18:50:03 +0000 (0:00:00.056)       0:01:19.755 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72
Thursday 21 July 2022  18:50:03 +0000 (0:00:00.051)       0:01:19.806 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78
Thursday 21 July 2022  18:50:03 +0000 (0:00:00.036)       0:01:19.842 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84
Thursday 21 July 2022  18:50:03 +0000 (0:00:00.038)       0:01:19.881 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.045)       0:01:19.926 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [get information about RAID] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.040)       0:01:19.967 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.040)       0:01:20.007 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.040)       0:01:20.048 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.036)       0:01:20.084 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.039)       0:01:20.124 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.035)       0:01:20.159 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.036)       0:01:20.195 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the actual size of the volume] *************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.035)       0:01:20.231 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested size of the volume] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.026)       0:01:20.257 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.038)       0:01:20.296 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.036)       0:01:20.332 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.078)       0:01:20.410 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.039)       0:01:20.449 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.037)       0:01:20.487 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.036)       0:01:20.524 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.039)       0:01:20.563 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.097)       0:01:20.661 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.038)       0:01:20.699 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.035)       0:01:20.735 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.025)       0:01:20.760 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.026)       0:01:20.786 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check segment type] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.027)       0:01:20.814 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.022)       0:01:20.836 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested cache size] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.025)       0:01:20.862 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32
Thursday 21 July 2022  18:50:04 +0000 (0:00:00.025)       0:01:20.888 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.023)       0:01:20.912 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.022)       0:01:20.935 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.034)       0:01:20.970 ********* 

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.020)       0:01:20.990 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Test setting up disk volume will remove the partition create above] ******
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:176
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.038)       0:01:21.028 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.048)       0:01:21.077 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.036)       0:01:21.113 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.421)       0:01:21.535 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.065)       0:01:21.601 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.039)       0:01:21.640 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.037)       0:01:21.678 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.056)       0:01:21.734 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.023)       0:01:21.758 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.028)       0:01:21.787 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.039)       0:01:21.826 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": [
        {
            "disks": [
                "nvme1n1"
            ],
            "fs_create_options": "-F",
            "fs_type": "ext4",
            "mount_options": "rw,noatime,defaults",
            "mount_point": "/opt/test1",
            "name": "foo",
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.040)       0:01:21.867 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:50:05 +0000 (0:00:00.029)       0:01:21.896 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:50:06 +0000 (0:00:00.028)       0:01:21.924 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:50:06 +0000 (0:00:00.027)       0:01:21.952 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:50:06 +0000 (0:00:00.026)       0:01:21.979 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:50:06 +0000 (0:00:00.058)       0:01:22.037 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:50:06 +0000 (0:00:00.023)       0:01:22.061 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/nvme1n1p1",
            "fs_type": "ext4"
        },
        {
            "action": "destroy device",
            "device": "/dev/nvme1n1p1",
            "fs_type": null
        },
        {
            "action": "destroy format",
            "device": "/dev/nvme1n1",
            "fs_type": "disklabel"
        },
        {
            "action": "create format",
            "device": "/dev/nvme1n1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6",
            "state": "absent"
        },
        {
            "dump": 0,
            "fstype": "ext4",
            "opts": "rw,noatime,defaults",
            "passno": 0,
            "path": "/opt/test1",
            "src": "UUID=108ace4b-02eb-4de4-ae10-a165744cba4a",
            "state": "mounted"
        }
    ],
    "packages": [
        "xfsprogs",
        "e2fsprogs"
    ],
    "pools": [],
    "volumes": [
        {
            "_device": "/dev/nvme1n1",
            "_kernel_device": "/dev/nvme1n1",
            "_mount_id": "UUID=108ace4b-02eb-4de4-ae10-a165744cba4a",
            "_raw_device": "/dev/nvme1n1",
            "_raw_kernel_device": "/dev/nvme1n1",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "fs_create_options": "-F",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "ext4",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_options": "rw,noatime,defaults",
            "mount_passno": 0,
            "mount_point": "/opt/test1",
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "size": 10737418240,
            "state": "present",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:50:07 +0000 (0:00:01.634)       0:01:23.696 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:50:07 +0000 (0:00:00.036)       0:01:23.732 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:50:07 +0000 (0:00:00.021)       0:01:23.753 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/nvme1n1p1",
                "fs_type": "ext4"
            },
            {
                "action": "destroy device",
                "device": "/dev/nvme1n1p1",
                "fs_type": null
            },
            {
                "action": "destroy format",
                "device": "/dev/nvme1n1",
                "fs_type": "disklabel"
            },
            {
                "action": "create format",
                "device": "/dev/nvme1n1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6",
                "state": "absent"
            },
            {
                "dump": 0,
                "fstype": "ext4",
                "opts": "rw,noatime,defaults",
                "passno": 0,
                "path": "/opt/test1",
                "src": "UUID=108ace4b-02eb-4de4-ae10-a165744cba4a",
                "state": "mounted"
            }
        ],
        "packages": [
            "xfsprogs",
            "e2fsprogs"
        ],
        "pools": [],
        "volumes": [
            {
                "_device": "/dev/nvme1n1",
                "_kernel_device": "/dev/nvme1n1",
                "_mount_id": "UUID=108ace4b-02eb-4de4-ae10-a165744cba4a",
                "_raw_device": "/dev/nvme1n1",
                "_raw_kernel_device": "/dev/nvme1n1",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "-F",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "ext4",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_options": "rw,noatime,defaults",
                "mount_passno": 0,
                "mount_point": "/opt/test1",
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "size": 10737418240,
                "state": "present",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:50:07 +0000 (0:00:00.038)       0:01:23.792 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:50:07 +0000 (0:00:00.040)       0:01:23.833 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": [
            {
                "_device": "/dev/nvme1n1",
                "_kernel_device": "/dev/nvme1n1",
                "_mount_id": "UUID=108ace4b-02eb-4de4-ae10-a165744cba4a",
                "_raw_device": "/dev/nvme1n1",
                "_raw_kernel_device": "/dev/nvme1n1",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "-F",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "ext4",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_options": "rw,noatime,defaults",
                "mount_passno": 0,
                "mount_point": "/opt/test1",
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "size": 10737418240,
                "state": "present",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:50:07 +0000 (0:00:00.040)       0:01:23.873 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': 'UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "UUID=9ce48a48-ec04-43d1-aaf5-dd298f4a75f6"
}

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:50:08 +0000 (0:00:00.336)       0:01:24.210 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:50:08 +0000 (0:00:00.447)       0:01:24.657 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': 'UUID=108ace4b-02eb-4de4-ae10-a165744cba4a', 'dump': 0, 'passno': 0, 'fstype': 'ext4', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'rw,noatime,defaults'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "dump": 0,
        "fstype": "ext4",
        "opts": "rw,noatime,defaults",
        "passno": 0,
        "path": "/opt/test1",
        "src": "UUID=108ace4b-02eb-4de4-ae10-a165744cba4a",
        "state": "mounted"
    },
    "name": "/opt/test1",
    "opts": "rw,noatime,defaults",
    "passno": "0",
    "src": "UUID=108ace4b-02eb-4de4-ae10-a165744cba4a"
}

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:50:09 +0000 (0:00:00.342)       0:01:25.000 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:50:09 +0000 (0:00:00.456)       0:01:25.456 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429333.3914723,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:50:09 +0000 (0:00:00.331)       0:01:25.787 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:50:09 +0000 (0:00:00.022)       0:01:25.810 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:189
Thursday 21 July 2022  18:50:10 +0000 (0:00:00.850)       0:01:26.661 ********* 
included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2

TASK [Print out pool information] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1
Thursday 21 July 2022  18:50:10 +0000 (0:00:00.041)       0:01:26.703 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Print out volume information] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6
Thursday 21 July 2022  18:50:10 +0000 (0:00:00.037)       0:01:26.740 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "_storage_volumes_list": [
        {
            "_device": "/dev/nvme1n1",
            "_kernel_device": "/dev/nvme1n1",
            "_mount_id": "UUID=108ace4b-02eb-4de4-ae10-a165744cba4a",
            "_raw_device": "/dev/nvme1n1",
            "_raw_kernel_device": "/dev/nvme1n1",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "fs_create_options": "-F",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "ext4",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_options": "rw,noatime,defaults",
            "mount_passno": 0,
            "mount_point": "/opt/test1",
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "size": 10737418240,
            "state": "present",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [Collect info about the volumes.] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14
Thursday 21 July 2022  18:50:10 +0000 (0:00:00.052)       0:01:26.792 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "info": {
        "/dev/fd0": {
            "fstype": "",
            "label": "",
            "name": "/dev/fd0",
            "size": "4K",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme0n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme0n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/nvme1n1": {
            "fstype": "ext4",
            "label": "",
            "name": "/dev/nvme1n1",
            "size": "10G",
            "type": "disk",
            "uuid": "108ace4b-02eb-4de4-ae10-a165744cba4a"
        },
        "/dev/nvme2n1": {
            "fstype": "",
            "label": "",
            "name": "/dev/nvme2n1",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sda": {
            "fstype": "",
            "label": "",
            "name": "/dev/sda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/sdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/sr0": {
            "fstype": "iso9660",
            "label": "cidata",
            "name": "/dev/sr0",
            "size": "364K",
            "type": "rom",
            "uuid": "2022-07-21-18-48-31-00"
        },
        "/dev/vda": {
            "fstype": "",
            "label": "",
            "name": "/dev/vda",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vda1": {
            "fstype": "xfs",
            "label": "",
            "name": "/dev/vda1",
            "size": "10G",
            "type": "partition",
            "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053"
        },
        "/dev/vdb": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdb",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdc": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdc",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        },
        "/dev/vdd": {
            "fstype": "",
            "label": "",
            "name": "/dev/vdd",
            "size": "10G",
            "type": "disk",
            "uuid": ""
        }
    }
}

TASK [Read the /etc/fstab file for volume existence] ***************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19
Thursday 21 July 2022  18:50:12 +0000 (0:00:01.319)       0:01:28.112 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/fstab"
    ],
    "delta": "0:00:00.003816",
    "end": "2022-07-21 14:50:12.429668",
    "rc": 0,
    "start": "2022-07-21 14:50:12.425852"
}

STDOUT:


#
# /etc/fstab
# Created by anaconda on Tue Jul 19 03:15:15 2022
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 /                       xfs     defaults        0 0
UUID=108ace4b-02eb-4de4-ae10-a165744cba4a /opt/test1 ext4 rw,noatime,defaults 0 0

TASK [Read the /etc/crypttab file] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24
Thursday 21 July 2022  18:50:12 +0000 (0:00:00.371)       0:01:28.484 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "cmd": [
        "cat",
        "/etc/crypttab"
    ],
    "delta": "0:00:00.004281",
    "end": "2022-07-21 14:50:12.736187",
    "failed_when_result": false,
    "rc": 0,
    "start": "2022-07-21 14:50:12.731906"
}

TASK [Verify the volumes listed in storage_pools were correctly managed] *******
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33
Thursday 21 July 2022  18:50:12 +0000 (0:00:00.346)       0:01:28.830 ********* 

TASK [Verify the volumes with no pool were correctly managed] ******************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43
Thursday 21 July 2022  18:50:12 +0000 (0:00:00.021)       0:01:28.852 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/nvme1n1', 'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'ext4', 'mount_options': 'rw,noatime,defaults', '_device': '/dev/nvme1n1', 'size': 10737418240, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/nvme1n1', 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'type': 'disk', 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': 'UUID=108ace4b-02eb-4de4-ae10-a165744cba4a', 'raid_spare_count': None, 'name': 'foo', '_raw_kernel_device': '/dev/nvme1n1', 'cache_mode': None, 'cache_devices': [], 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': None, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'thin_pool_size': None, 'fs_create_options': '-F'})

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.057)       0:01:28.909 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": true,
        "_storage_volume_tests": [
            "mount",
            "fstab",
            "fs",
            "device",
            "encryption",
            "md",
            "size",
            "cache"
        ]
    },
    "changed": false
}

TASK [include_tasks] ***********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.049)       0:01:28.959 ********* 
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size)
included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache)

TASK [Get expected mount device based on device type] **************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.073)       0:01:29.033 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_device_path": "/dev/nvme1n1"
    },
    "changed": false
}

TASK [Set some facts] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.038)       0:01:29.071 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": [
            {
                "block_available": 2403135,
                "block_size": 4096,
                "block_total": 2547525,
                "block_used": 144390,
                "device": "/dev/nvme1n1",
                "fstype": "ext4",
                "inode_available": 655349,
                "inode_total": 655360,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,noatime,data=ordered",
                "size_available": 9843240960,
                "size_total": 10434662400,
                "uuid": "108ace4b-02eb-4de4-ae10-a165744cba4a"
            }
        ],
        "storage_test_mount_expected_match_count": "1",
        "storage_test_mount_point_matches": [
            {
                "block_available": 2403135,
                "block_size": 4096,
                "block_total": 2547525,
                "block_used": 144390,
                "device": "/dev/nvme1n1",
                "fstype": "ext4",
                "inode_available": 655349,
                "inode_total": 655360,
                "inode_used": 11,
                "mount": "/opt/test1",
                "options": "rw,seclabel,noatime,data=ordered",
                "size_available": 9843240960,
                "size_total": 10434662400,
                "uuid": "108ace4b-02eb-4de4-ae10-a165744cba4a"
            }
        ],
        "storage_test_swap_expected_matches": "0"
    },
    "changed": false
}

TASK [Verify the current mount state by device] ********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.053)       0:01:29.125 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the current mount state by mount point] ***************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.054)       0:01:29.179 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the mount fs type] ************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.051)       0:01:29.231 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [command] *****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.050)       0:01:29.282 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Gather swap info] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.025)       0:01:29.307 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify swap status] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.023)       0:01:29.331 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Unset facts] *************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.023)       0:01:29.354 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_mount_device_matches": null,
        "storage_test_mount_expected_match_count": null,
        "storage_test_mount_point_matches": null,
        "storage_test_swap_expected_matches": null,
        "storage_test_swaps": null,
        "storage_test_sys_node": null
    },
    "changed": false
}

TASK [Set some variables for fstab checking] ***********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.037)       0:01:29.392 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": "1",
        "storage_test_fstab_expected_mount_options_matches": "1",
        "storage_test_fstab_expected_mount_point_matches": "1",
        "storage_test_fstab_id_matches": [
            "UUID=108ace4b-02eb-4de4-ae10-a165744cba4a "
        ],
        "storage_test_fstab_mount_options_matches": [
            " /opt/test1 ext4 rw,noatime,defaults "
        ],
        "storage_test_fstab_mount_point_matches": [
            " /opt/test1 "
        ]
    },
    "changed": false
}

TASK [Verify that the device identifier appears in /etc/fstab] *****************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.065)       0:01:29.457 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the fstab mount point] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.048)       0:01:29.506 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify mount_options] ****************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.049)       0:01:29.556 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Clean up variables] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.053)       0:01:29.609 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_fstab_expected_id_matches": null,
        "storage_test_fstab_expected_mount_options_matches": null,
        "storage_test_fstab_expected_mount_point_matches": null,
        "storage_test_fstab_id_matches": null,
        "storage_test_fstab_mount_options_matches": null,
        "storage_test_fstab_mount_point_matches": null
    },
    "changed": false
}

TASK [Verify fs type] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.035)       0:01:29.645 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify fs label] *********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.040)       0:01:29.685 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [See whether the device node is present] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4
Thursday 21 July 2022  18:50:13 +0000 (0:00:00.040)       0:01:29.725 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429407.6144722,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "ctime": 1658429407.6144722,
        "dev": 5,
        "device_type": 66305,
        "executable": false,
        "exists": true,
        "gid": 6,
        "gr_name": "disk",
        "inode": 11266,
        "isblk": true,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": false,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/blockdevice",
        "mode": "0660",
        "mtime": 1658429407.6144722,
        "nlink": 1,
        "path": "/dev/nvme1n1",
        "pw_name": "root",
        "readable": true,
        "rgrp": true,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": null,
        "wgrp": true,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [Verify the presence/absence of the device node] **************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10
Thursday 21 July 2022  18:50:14 +0000 (0:00:00.317)       0:01:30.043 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about this volume] *********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15
Thursday 21 July 2022  18:50:14 +0000 (0:00:00.036)       0:01:30.080 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [(1/2) Process volume type (set initial value)] ***************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21
Thursday 21 July 2022  18:50:14 +0000 (0:00:00.041)       0:01:30.121 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "st_volume_type": "disk"
    },
    "changed": false
}

TASK [(2/2) Process volume type (get RAID value)] ******************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25
Thursday 21 July 2022  18:50:14 +0000 (0:00:00.037)       0:01:30.158 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the volume's device type] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30
Thursday 21 July 2022  18:50:14 +0000 (0:00:00.023)       0:01:30.182 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Stat the LUKS device, if encrypted] **************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3
Thursday 21 July 2022  18:50:14 +0000 (0:00:00.038)       0:01:30.220 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Ensure cryptsetup is present] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10
Thursday 21 July 2022  18:50:14 +0000 (0:00:00.021)       0:01:30.242 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "rc": 0,
    "results": [
        "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed"
    ]
}

TASK [Collect LUKS info for this volume] ***************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15
Thursday 21 July 2022  18:50:14 +0000 (0:00:00.530)       0:01:30.773 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the presence/absence of the LUKS device node] *********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21
Thursday 21 July 2022  18:50:14 +0000 (0:00:00.024)       0:01:30.797 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify that the raw device is the same as the device if not encrypted] ***
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27
Thursday 21 July 2022  18:50:14 +0000 (0:00:00.023)       0:01:30.821 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Make sure we got info about the LUKS volume if encrypted] ****************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33
Thursday 21 July 2022  18:50:14 +0000 (0:00:00.051)       0:01:30.873 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Verify the LUKS volume's device type if encrypted] ***********************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39
Thursday 21 July 2022  18:50:14 +0000 (0:00:00.023)       0:01:30.897 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS version] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.021)       0:01:30.918 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS key size] *****************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.021)       0:01:30.939 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check LUKS cipher] *******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.022)       0:01:30.962 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.023)       0:01:30.985 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": [],
        "_storage_test_expected_crypttab_entries": "0",
        "_storage_test_expected_crypttab_key_file": "-"
    },
    "changed": false
}

TASK [Check for /etc/crypttab entry] *******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.086)       0:01:31.072 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Validate the format of the crypttab entry] *******************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.118)       0:01:31.190 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check backing device of crypttab entry] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.037)       0:01:31.228 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check key file of crypttab entry] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.035)       0:01:31.263 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.035)       0:01:31.299 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_crypttab_entries": null,
        "_storage_test_expected_crypttab_entries": null,
        "_storage_test_expected_crypttab_key_file": null
    },
    "changed": false
}

TASK [get information about RAID] **********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.036)       0:01:31.336 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.036)       0:01:31.372 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.036)       0:01:31.409 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.038)       0:01:31.447 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID active devices count] *****************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.038)       0:01:31.485 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID spare devices count] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.039)       0:01:31.525 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check RAID metadata version] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.038)       0:01:31.563 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the actual size of the volume] *************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.037)       0:01:31.601 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested size of the volume] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.025)       0:01:31.627 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Establish base value for expected size] **********************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.035)       0:01:31.662 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.035)       0:01:31.697 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.034)       0:01:31.731 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.035)       0:01:31.767 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Get the size of parent/pool device] **************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.036)       0:01:31.804 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.035)       0:01:31.840 ********* 
skipping: [/cache/rhel-7.qcow2] => {}

TASK [Calculate the expected size based on pool size and percentage value] *****
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39
Thursday 21 July 2022  18:50:15 +0000 (0:00:00.036)       0:01:31.876 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.039)       0:01:31.916 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_actual_size": {
        "changed": false,
        "skip_reason": "Conditional result was False",
        "skipped": true
    }
}

TASK [debug] *******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.037)       0:01:31.954 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_test_expected_size": "4294967296"
}

TASK [assert] ******************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.037)       0:01:31.992 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Get information about the LV] ********************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.024)       0:01:32.016 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.026)       0:01:32.042 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [check segment type] ******************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.024)       0:01:32.066 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.024)       0:01:32.090 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [parse the requested cache size] ******************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.024)       0:01:32.115 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [set_fact] ****************************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.025)       0:01:32.140 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Check cache size] ********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.024)       0:01:32.164 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [Clean up facts] **********************************************************
task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.023)       0:01:32.187 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_test_volume_present": null
    },
    "changed": false
}

TASK [Clean up variable namespace] *********************************************
task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.035)       0:01:32.223 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_test_blkinfo": null,
        "storage_test_crypttab": null,
        "storage_test_fstab": null
    },
    "changed": false
}

TASK [Remove the disk volume created above] ************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:193
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.036)       0:01:32.260 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.053)       0:01:32.313 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.036)       0:01:32.349 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.417)       0:01:32.767 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.064)       0:01:32.832 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:50:16 +0000 (0:00:00.035)       0:01:32.867 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:50:17 +0000 (0:00:00.079)       0:01:32.947 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:50:17 +0000 (0:00:00.057)       0:01:33.004 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:50:17 +0000 (0:00:00.021)       0:01:33.026 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:50:17 +0000 (0:00:00.068)       0:01:33.095 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:50:17 +0000 (0:00:00.036)       0:01:33.132 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": [
        {
            "disks": [
                "nvme1n1"
            ],
            "name": "foo",
            "state": "absent",
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:50:17 +0000 (0:00:00.050)       0:01:33.182 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:50:17 +0000 (0:00:00.029)       0:01:33.211 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:50:17 +0000 (0:00:00.030)       0:01:33.242 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:50:17 +0000 (0:00:00.025)       0:01:33.268 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:50:17 +0000 (0:00:00.028)       0:01:33.297 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:50:17 +0000 (0:00:00.056)       0:01:33.353 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:50:17 +0000 (0:00:00.020)       0:01:33.374 ********* 
changed: [/cache/rhel-7.qcow2] => {
    "actions": [
        {
            "action": "destroy format",
            "device": "/dev/nvme1n1",
            "fs_type": "ext4"
        }
    ],
    "changed": true,
    "crypts": [],
    "leaves": [
        "/dev/sr0",
        "/dev/vda1",
        "/dev/sda",
        "/dev/sdb",
        "/dev/sdc",
        "/dev/nvme0n1",
        "/dev/nvme1n1",
        "/dev/nvme2n1",
        "/dev/vdb",
        "/dev/vdc",
        "/dev/vdd"
    ],
    "mounts": [
        {
            "fstype": "ext4",
            "path": "/opt/test1",
            "src": "UUID=108ace4b-02eb-4de4-ae10-a165744cba4a",
            "state": "absent"
        }
    ],
    "packages": [
        "xfsprogs"
    ],
    "pools": [],
    "volumes": [
        {
            "_device": "/dev/nvme1n1",
            "_mount_id": "UUID=108ace4b-02eb-4de4-ae10-a165744cba4a",
            "_raw_device": "/dev/nvme1n1",
            "cache_devices": [],
            "cache_mode": null,
            "cache_size": 0,
            "cached": false,
            "compression": null,
            "deduplication": null,
            "disks": [
                "nvme1n1"
            ],
            "encryption": false,
            "encryption_cipher": null,
            "encryption_key": null,
            "encryption_key_size": null,
            "encryption_luks_version": null,
            "encryption_password": null,
            "fs_create_options": "",
            "fs_label": "",
            "fs_overwrite_existing": true,
            "fs_type": "ext4",
            "mount_check": 0,
            "mount_device_identifier": "uuid",
            "mount_options": "defaults",
            "mount_passno": 0,
            "mount_point": null,
            "name": "foo",
            "raid_chunk_size": null,
            "raid_device_count": null,
            "raid_level": null,
            "raid_metadata_version": null,
            "raid_spare_count": null,
            "size": 10737418240,
            "state": "absent",
            "thin": null,
            "thin_pool_name": null,
            "thin_pool_size": null,
            "type": "disk",
            "vdo_pool_size": null
        }
    ]
}

TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78
Thursday 21 July 2022  18:50:18 +0000 (0:00:01.049)       0:01:34.423 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:50:18 +0000 (0:00:00.040)       0:01:34.463 ********* 

TASK [fedora.linux_system_roles.storage : show blivet_output] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96
Thursday 21 July 2022  18:50:18 +0000 (0:00:00.020)       0:01:34.484 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "blivet_output": {
        "actions": [
            {
                "action": "destroy format",
                "device": "/dev/nvme1n1",
                "fs_type": "ext4"
            }
        ],
        "changed": true,
        "crypts": [],
        "failed": false,
        "leaves": [
            "/dev/sr0",
            "/dev/vda1",
            "/dev/sda",
            "/dev/sdb",
            "/dev/sdc",
            "/dev/nvme0n1",
            "/dev/nvme1n1",
            "/dev/nvme2n1",
            "/dev/vdb",
            "/dev/vdc",
            "/dev/vdd"
        ],
        "mounts": [
            {
                "fstype": "ext4",
                "path": "/opt/test1",
                "src": "UUID=108ace4b-02eb-4de4-ae10-a165744cba4a",
                "state": "absent"
            }
        ],
        "packages": [
            "xfsprogs"
        ],
        "pools": [],
        "volumes": [
            {
                "_device": "/dev/nvme1n1",
                "_mount_id": "UUID=108ace4b-02eb-4de4-ae10-a165744cba4a",
                "_raw_device": "/dev/nvme1n1",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "ext4",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_options": "defaults",
                "mount_passno": 0,
                "mount_point": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "size": 10737418240,
                "state": "absent",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    }
}

TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101
Thursday 21 July 2022  18:50:18 +0000 (0:00:00.036)       0:01:34.520 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105
Thursday 21 July 2022  18:50:18 +0000 (0:00:00.035)       0:01:34.556 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": [
            {
                "_device": "/dev/nvme1n1",
                "_mount_id": "UUID=108ace4b-02eb-4de4-ae10-a165744cba4a",
                "_raw_device": "/dev/nvme1n1",
                "cache_devices": [],
                "cache_mode": null,
                "cache_size": 0,
                "cached": false,
                "compression": null,
                "deduplication": null,
                "disks": [
                    "nvme1n1"
                ],
                "encryption": false,
                "encryption_cipher": null,
                "encryption_key": null,
                "encryption_key_size": null,
                "encryption_luks_version": null,
                "encryption_password": null,
                "fs_create_options": "",
                "fs_label": "",
                "fs_overwrite_existing": true,
                "fs_type": "ext4",
                "mount_check": 0,
                "mount_device_identifier": "uuid",
                "mount_options": "defaults",
                "mount_passno": 0,
                "mount_point": null,
                "name": "foo",
                "raid_chunk_size": null,
                "raid_device_count": null,
                "raid_level": null,
                "raid_metadata_version": null,
                "raid_spare_count": null,
                "size": 10737418240,
                "state": "absent",
                "thin": null,
                "thin_pool_name": null,
                "thin_pool_size": null,
                "type": "disk",
                "vdo_pool_size": null
            }
        ]
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : remove obsolete mounts] **************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121
Thursday 21 July 2022  18:50:18 +0000 (0:00:00.037)       0:01:34.593 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
changed: [/cache/rhel-7.qcow2] => (item={'src': 'UUID=108ace4b-02eb-4de4-ae10-a165744cba4a', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'ext4'}) => {
    "ansible_loop_var": "mount_info",
    "backup_file": "",
    "boot": "yes",
    "changed": true,
    "dump": "0",
    "fstab": "/etc/fstab",
    "fstype": "ext4",
    "mount_info": {
        "fstype": "ext4",
        "path": "/opt/test1",
        "src": "UUID=108ace4b-02eb-4de4-ae10-a165744cba4a",
        "state": "absent"
    },
    "name": "/opt/test1",
    "opts": "defaults",
    "passno": "0",
    "src": "UUID=108ace4b-02eb-4de4-ae10-a165744cba4a"
}

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132
Thursday 21 July 2022  18:50:19 +0000 (0:00:00.337)       0:01:34.931 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : set up new/current mounts] ***********
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137
Thursday 21 July 2022  18:50:19 +0000 (0:00:00.439)       0:01:35.370 ********* 

TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148
Thursday 21 July 2022  18:50:19 +0000 (0:00:00.037)       0:01:35.408 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "name": null,
    "status": {}
}

TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156
Thursday 21 July 2022  18:50:19 +0000 (0:00:00.416)       0:01:35.824 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "stat": {
        "atime": 1658429333.3914723,
        "attr_flags": "",
        "attributes": [],
        "block_size": 4096,
        "blocks": 0,
        "charset": "binary",
        "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
        "ctime": 1658201031.524,
        "dev": 64769,
        "device_type": 0,
        "executable": false,
        "exists": true,
        "gid": 0,
        "gr_name": "root",
        "inode": 70,
        "isblk": false,
        "ischr": false,
        "isdir": false,
        "isfifo": false,
        "isgid": false,
        "islnk": false,
        "isreg": true,
        "issock": false,
        "isuid": false,
        "mimetype": "inode/x-empty",
        "mode": "0600",
        "mtime": 1658200515.884,
        "nlink": 1,
        "path": "/etc/crypttab",
        "pw_name": "root",
        "readable": true,
        "rgrp": false,
        "roth": false,
        "rusr": true,
        "size": 0,
        "uid": 0,
        "version": "18446744071677828413",
        "wgrp": false,
        "woth": false,
        "writeable": true,
        "wusr": true,
        "xgrp": false,
        "xoth": false,
        "xusr": false
    }
}

TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161
Thursday 21 July 2022  18:50:20 +0000 (0:00:00.289)       0:01:36.114 ********* 

TASK [fedora.linux_system_roles.storage : Update facts] ************************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183
Thursday 21 July 2022  18:50:20 +0000 (0:00:00.021)       0:01:36.136 ********* 
ok: [/cache/rhel-7.qcow2]
META: role_complete for /cache/rhel-7.qcow2

TASK [Try to mount swap filesystem to "/opt/test1"] ****************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:205
Thursday 21 July 2022  18:50:21 +0000 (0:00:00.806)       0:01:36.943 ********* 

TASK [fedora.linux_system_roles.storage : set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2
Thursday 21 July 2022  18:50:21 +0000 (0:00:00.037)       0:01:36.981 ********* 
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2
Thursday 21 July 2022  18:50:21 +0000 (0:00:00.033)       0:01:37.014 ********* 
ok: [/cache/rhel-7.qcow2]

TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7
Thursday 21 July 2022  18:50:21 +0000 (0:00:00.442)       0:01:37.456 ********* 
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat.yml",
    "skip_reason": "Conditional result was False"
}
ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => {
    "ansible_facts": {
        "__storage_blivet_diskvolume_mkfs_option_map": {
            "ext2": "-F",
            "ext3": "-F",
            "ext4": "-F"
        },
        "blivet_package_list": [
            "python-enum34",
            "python-blivet3",
            "libblockdev-crypto",
            "libblockdev-dm",
            "libblockdev-lvm",
            "libblockdev-mdraid",
            "libblockdev-swap"
        ]
    },
    "ansible_included_var_files": [
        "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml"
    ],
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.yml"
}
skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml)  => {
    "ansible_loop_var": "item",
    "changed": false,
    "item": "RedHat_7.9.yml",
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5
Thursday 21 July 2022  18:50:21 +0000 (0:00:00.058)       0:01:37.515 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_pools_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9
Thursday 21 July 2022  18:50:21 +0000 (0:00:00.033)       0:01:37.548 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "_storage_volumes_list": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13
Thursday 21 July 2022  18:50:21 +0000 (0:00:00.031)       0:01:37.580 ********* 
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount
included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2

TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2
Thursday 21 July 2022  18:50:21 +0000 (0:00:00.052)       0:01:37.632 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure blivet is available] *******
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7
Thursday 21 July 2022  18:50:21 +0000 (0:00:00.020)       0:01:37.653 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : show storage_pools] ******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13
Thursday 21 July 2022  18:50:21 +0000 (0:00:00.031)       0:01:37.684 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined"
}

TASK [fedora.linux_system_roles.storage : show storage_volumes] ****************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18
Thursday 21 July 2022  18:50:21 +0000 (0:00:00.034)       0:01:37.719 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "storage_volumes": [
        {
            "disks": [
                "nvme1n1"
            ],
            "fs_type": "swap",
            "mount_point": "/opt/test1",
            "name": "test1",
            "type": "disk"
        }
    ]
}

TASK [fedora.linux_system_roles.storage : get required packages] ***************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23
Thursday 21 July 2022  18:50:21 +0000 (0:00:00.048)       0:01:37.768 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35
Thursday 21 July 2022  18:50:21 +0000 (0:00:00.025)       0:01:37.794 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : make sure required packages are installed] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41
Thursday 21 July 2022  18:50:21 +0000 (0:00:00.025)       0:01:37.819 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : get service facts] *******************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47
Thursday 21 July 2022  18:50:21 +0000 (0:00:00.025)       0:01:37.845 ********* 
skipping: [/cache/rhel-7.qcow2] => {
    "changed": false,
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] *****
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53
Thursday 21 July 2022  18:50:21 +0000 (0:00:00.025)       0:01:37.871 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "ansible_facts": {
        "storage_cryptsetup_services": []
    },
    "changed": false
}

TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58
Thursday 21 July 2022  18:50:22 +0000 (0:00:00.053)       0:01:37.924 ********* 

TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64
Thursday 21 July 2022  18:50:22 +0000 (0:00:00.020)       0:01:37.944 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "actions": [],
    "changed": false,
    "crypts": [],
    "leaves": [],
    "mounts": [],
    "packages": [],
    "pools": [],
    "volumes": []
}

MSG:

volume 'test1' has a mount point but no mountable file system

TASK [fedora.linux_system_roles.storage : failed message] **********************
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:86
Thursday 21 July 2022  18:50:22 +0000 (0:00:00.830)       0:01:38.775 ********* 
fatal: [/cache/rhel-7.qcow2]: FAILED! => {
    "changed": false
}

MSG:

{'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': False, 'pools': [], 'volumes': [{'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'swap', 'mount_options': 'defaults', 'size': 10737418240, 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'type': 'disk', 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, 'raid_spare_count': None, 'name': 'test1', 'cache_mode': None, 'cache_devices': [], 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': None, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'thin_pool_size': None, 'fs_create_options': ''}], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': [], 'msg': "volume 'test1' has a mount point but no mountable file system", '_ansible_no_log': False}

TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] ***
task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90
Thursday 21 July 2022  18:50:22 +0000 (0:00:00.042)       0:01:38.818 ********* 

TASK [Check that we failed in the role] ****************************************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:221
Thursday 21 July 2022  18:50:22 +0000 (0:00:00.021)       0:01:38.839 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Verify the output when mount swap filesystem to "/opt/test1"] ************
task path: /tmp/tmptomayb7j/tests/storage/tests_misc.yml:227
Thursday 21 July 2022  18:50:22 +0000 (0:00:00.038)       0:01:38.877 ********* 
ok: [/cache/rhel-7.qcow2] => {
    "changed": false
}

MSG:

All assertions passed
META: ran handlers
META: ran handlers

PLAY RECAP *********************************************************************
/cache/rhel-7.qcow2        : ok=481  changed=19   unreachable=0    failed=3    skipped=367  rescued=3    ignored=0   

Thursday 21 July 2022  18:50:23 +0000 (0:00:00.062)       0:01:38.940 ********* 
=============================================================================== 
fedora.linux_system_roles.storage : make sure blivet is available ------- 8.62s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 
fedora.linux_system_roles.storage : Update facts ------------------------ 1.81s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.66s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.63s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.58s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.56s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.54s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.52s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
Ensure cryptsetup is present -------------------------------------------- 1.45s
/tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 -----------
Collect info about the volumes. ----------------------------------------- 1.42s
/tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14 ---------------------
set up internal repositories -------------------------------------------- 1.38s
/cache/rhel-7_setup.yml:5 -----------------------------------------------------
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.38s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.36s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
Collect info about the volumes. ----------------------------------------- 1.32s
/tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14 ---------------------
Collect info about the volumes. ----------------------------------------- 1.32s
/tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14 ---------------------
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.21s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
fedora.linux_system_roles.storage : get service facts ------------------- 1.13s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 
fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.05s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 
Gathering Facts --------------------------------------------------------- 0.98s
/tmp/tmptomayb7j/tests/storage/tests_misc.yml:2 -------------------------------
fedora.linux_system_roles.storage : Update facts ------------------------ 0.85s
/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183