ansible-playbook 2.9.27 config file = /etc/ansible/ansible.cfg configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 2.7.5 (default, Nov 14 2023, 16:14:06) [GCC 4.8.5 20150623 (Red Hat 4.8.5-44)] Using /etc/ansible/ansible.cfg as config file [WARNING]: running playbook inside collection fedora.linux_system_roles Skipping callback 'actionable', as we already have a stdout callback. Skipping callback 'counter_enabled', as we already have a stdout callback. Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'full_skip', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'null', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. Skipping callback 'selective', as we already have a stdout callback. Skipping callback 'skippy', as we already have a stdout callback. Skipping callback 'stderr', as we already have a stdout callback. Skipping callback 'unixy', as we already have a stdout callback. Skipping callback 'yaml', as we already have a stdout callback. PLAYBOOK: tests_set_thin_basic.yml ********************************************* 1 plays in /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tests_set_thin_basic.yml PLAY [Snapshot a set of logical volumes across different volume groups] ******** TASK [Gathering Facts] ********************************************************* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tests_set_thin_basic.yml:2 Monday 11 November 2024 19:21:17 -0500 (0:00:00.026) 0:00:00.026 ******* ok: [managed-node3] META: ran handlers TASK [Setup] ******************************************************************* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tests_set_thin_basic.yml:47 Monday 11 November 2024 19:21:18 -0500 (0:00:01.045) 0:00:01.071 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tasks/setup.yml for managed-node3 TASK [Check if system is ostree] *********************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tasks/setup.yml:10 Monday 11 November 2024 19:21:18 -0500 (0:00:00.094) 0:00:01.166 ******* ok: [managed-node3] => { "changed": false, "stat": { "exists": false } } TASK [Set mount parent] ******************************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tasks/setup.yml:15 Monday 11 November 2024 19:21:18 -0500 (0:00:00.423) 0:00:01.590 ******* ok: [managed-node3] => { "ansible_facts": { "test_mnt_parent": "/mnt" }, "changed": false } TASK [Run the storage role install base packages] ****************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tasks/setup.yml:19 Monday 11 November 2024 19:21:18 -0500 (0:00:00.049) 0:00:01.639 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Monday 11 November 2024 19:21:18 -0500 (0:00:00.070) 0:00:01.709 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Monday 11 November 2024 19:21:18 -0500 (0:00:00.044) 0:00:01.753 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Monday 11 November 2024 19:21:18 -0500 (0:00:00.049) 0:00:01.802 ******* skipping: [managed-node3] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node3] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node3] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed-node3] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Monday 11 November 2024 19:21:19 -0500 (0:00:00.086) 0:00:01.889 ******* ok: [managed-node3] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Monday 11 November 2024 19:21:19 -0500 (0:00:00.310) 0:00:02.199 ******* ok: [managed-node3] => { "ansible_facts": { "__storage_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Monday 11 November 2024 19:21:19 -0500 (0:00:00.052) 0:00:02.251 ******* ok: [managed-node3] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Monday 11 November 2024 19:21:19 -0500 (0:00:00.024) 0:00:02.276 ******* ok: [managed-node3] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Monday 11 November 2024 19:21:19 -0500 (0:00:00.024) 0:00:02.300 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Monday 11 November 2024 19:21:19 -0500 (0:00:00.102) 0:00:02.403 ******* ok: [managed-node3] => { "changed": false, "rc": 0, "results": [ "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed" ] } lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34 TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Monday 11 November 2024 19:21:23 -0500 (0:00:04.125) 0:00:06.529 ******* ok: [managed-node3] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Monday 11 November 2024 19:21:23 -0500 (0:00:00.067) 0:00:06.596 ******* ok: [managed-node3] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Monday 11 November 2024 19:21:23 -0500 (0:00:00.059) 0:00:06.656 ******* ok: [managed-node3] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Monday 11 November 2024 19:21:24 -0500 (0:00:00.751) 0:00:07.408 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node3 TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Monday 11 November 2024 19:21:24 -0500 (0:00:00.181) 0:00:07.589 ******* TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Monday 11 November 2024 19:21:24 -0500 (0:00:00.071) 0:00:07.661 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************ task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19 Monday 11 November 2024 19:21:24 -0500 (0:00:00.084) 0:00:07.745 ******* TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Monday 11 November 2024 19:21:24 -0500 (0:00:00.052) 0:00:07.798 ******* ok: [managed-node3] => { "changed": false, "rc": 0, "results": [ "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed" ] } lsrpackages: kpartx TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Monday 11 November 2024 19:21:25 -0500 (0:00:00.843) 0:00:08.641 ******* ok: [managed-node3] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "brandbot.service": { "name": "brandbot.service", "source": "systemd", "state": "inactive", "status": "static" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-shell.service": { "name": "console-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.import1.service": { "name": "dbus-org.freedesktop.import1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.machine1.service": { "name": "dbus-org.freedesktop.machine1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "running", "status": "static" }, "dmraid-activation.service": { "name": "dmraid-activation.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-lvmetad.service": { "name": "lvm2-lvmetad.service", "source": "systemd", "state": "running", "status": "static" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "lvm2-pvscan@8:0.service": { "name": "lvm2-pvscan@8:0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:112.service": { "name": "lvm2-pvscan@8:112.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:128.service": { "name": "lvm2-pvscan@8:128.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:144.service": { "name": "lvm2-pvscan@8:144.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:16.service": { "name": "lvm2-pvscan@8:16.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:32.service": { "name": "lvm2-pvscan@8:32.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:48.service": { "name": "lvm2-pvscan@8:48.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:64.service": { "name": "lvm2-pvscan@8:64.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:80.service": { "name": "lvm2-pvscan@8:80.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:96.service": { "name": "lvm2-pvscan@8:96.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "netconsole": { "name": "netconsole", "source": "sysv", "state": "stopped", "status": "disabled" }, "network": { "name": "network", "source": "sysv", "state": "running", "status": "enabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-config.service": { "name": "nfs-config.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-idmap.service": { "name": "nfs-idmap.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-lock.service": { "name": "nfs-lock.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-rquotad.service": { "name": "nfs-rquotad.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-secure.service": { "name": "nfs-secure.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs.service": { "name": "nfs.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfslock.service": { "name": "nfslock.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "postfix.service": { "name": "postfix.service", "source": "systemd", "state": "running", "status": "enabled" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rhel-autorelabel-mark.service": { "name": "rhel-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-autorelabel.service": { "name": "rhel-autorelabel.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-configure.service": { "name": "rhel-configure.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-dmesg.service": { "name": "rhel-dmesg.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-domainname.service": { "name": "rhel-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-import-state.service": { "name": "rhel-import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-loadmodules.service": { "name": "rhel-loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-readonly.service": { "name": "rhel-readonly.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-rquotad.service": { "name": "rpc-rquotad.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpcgssd.service": { "name": "rpcgssd.service", "source": "systemd", "state": "inactive", "status": "static" }, "rpcidmapd.service": { "name": "rpcidmapd.service", "source": "systemd", "state": "inactive", "status": "static" }, "rsyncd.service": { "name": "rsyncd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyncd@.service": { "name": "rsyncd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-policy-migrate-local-changes@.service": { "name": "selinux-policy-migrate-local-changes@.service", "source": "systemd", "state": "unknown", "status": "static" }, "selinux-policy-migrate-local-changes@targeted.service": { "name": "selinux-policy-migrate-local-changes@targeted.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "unknown" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "static" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bootchart.service": { "name": "systemd-bootchart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-importd.service": { "name": "systemd-importd.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machined.service": { "name": "systemd-machined.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-nspawn@.service": { "name": "systemd-nspawn@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-readahead-collect.service": { "name": "systemd-readahead-collect.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-readahead-done.service": { "name": "systemd-readahead-done.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "systemd-readahead-drop.service": { "name": "systemd-readahead-drop.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "systemd-readahead-replay.service": { "name": "systemd-readahead-replay.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill@.service": { "name": "systemd-rfkill@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-shutdownd.service": { "name": "systemd-shutdownd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "wpa_supplicant.service": { "name": "wpa_supplicant.service", "source": "systemd", "state": "inactive", "status": "disabled" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Monday 11 November 2024 19:21:27 -0500 (0:00:01.179) 0:00:09.821 ******* ok: [managed-node3] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Monday 11 November 2024 19:21:27 -0500 (0:00:00.085) 0:00:09.906 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Monday 11 November 2024 19:21:27 -0500 (0:00:00.049) 0:00:09.955 ******* ok: [managed-node3] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Monday 11 November 2024 19:21:27 -0500 (0:00:00.623) 0:00:10.578 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Monday 11 November 2024 19:21:27 -0500 (0:00:00.041) 0:00:10.620 ******* ok: [managed-node3] => { "changed": false, "stat": { "atime": 1731370735.4243228, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "72884e3f126482c2d28276ff7c57744fa95eff91", "ctime": 1731370722.1001987, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263592, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1731370722.0991988, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1229, "uid": 0, "version": "18446744072116901260", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Monday 11 November 2024 19:21:28 -0500 (0:00:00.360) 0:00:10.980 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Monday 11 November 2024 19:21:28 -0500 (0:00:00.054) 0:00:11.035 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Monday 11 November 2024 19:21:28 -0500 (0:00:00.047) 0:00:11.082 ******* ok: [managed-node3] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Monday 11 November 2024 19:21:28 -0500 (0:00:00.046) 0:00:11.129 ******* ok: [managed-node3] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Monday 11 November 2024 19:21:28 -0500 (0:00:00.079) 0:00:11.208 ******* ok: [managed-node3] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Monday 11 November 2024 19:21:28 -0500 (0:00:00.047) 0:00:11.255 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Monday 11 November 2024 19:21:28 -0500 (0:00:00.049) 0:00:11.305 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Monday 11 November 2024 19:21:28 -0500 (0:00:00.049) 0:00:11.354 ******* TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Monday 11 November 2024 19:21:28 -0500 (0:00:00.051) 0:00:11.406 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Monday 11 November 2024 19:21:28 -0500 (0:00:00.050) 0:00:11.457 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Monday 11 November 2024 19:21:28 -0500 (0:00:00.056) 0:00:11.513 ******* ok: [managed-node3] => { "changed": false, "stat": { "atime": 1731370641.178454, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1718879272.062, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131079, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1718879026.308, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744072852913879", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Monday 11 November 2024 19:21:29 -0500 (0:00:00.469) 0:00:11.982 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Monday 11 November 2024 19:21:29 -0500 (0:00:00.120) 0:00:12.102 ******* ok: [managed-node3] TASK [Get unused disks] ******************************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tasks/setup.yml:25 Monday 11 November 2024 19:21:30 -0500 (0:00:00.914) 0:00:13.017 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml for managed-node3 TASK [Check if system is ostree] *********************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:5 Monday 11 November 2024 19:21:30 -0500 (0:00:00.077) 0:00:13.095 ******* ok: [managed-node3] => { "changed": false, "stat": { "exists": false } } TASK [Set flag to indicate system is ostree] *********************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:10 Monday 11 November 2024 19:21:30 -0500 (0:00:00.305) 0:00:13.400 ******* ok: [managed-node3] => { "ansible_facts": { "__snapshot_is_ostree": false }, "changed": false } TASK [Ensure test packages] **************************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:14 Monday 11 November 2024 19:21:30 -0500 (0:00:00.055) 0:00:13.455 ******* ok: [managed-node3] => { "changed": false, "rc": 0, "results": [ "util-linux-2.23.2-65.el7_9.1.x86_64 providing util-linux is already installed" ] } lsrpackages: util-linux TASK [Find unused disks in the system] ***************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:23 Monday 11 November 2024 19:21:31 -0500 (0:00:00.663) 0:00:14.118 ******* ok: [managed-node3] => { "changed": false, "disks": [ "sda", "sdb", "sdc", "sdd", "sde", "sdf", "sdg", "sdh", "sdi", "sdj" ], "info": [ "Line: NAME=\"/dev/sda\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdb\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdc\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdd\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sde\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdf\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdg\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdh\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdi\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdj\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdk\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdl\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/xvda\" TYPE=\"disk\" SIZE=\"268435456000\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/xvda1\" TYPE=\"part\" SIZE=\"268434390528\" FSTYPE=\"ext4\" LOG-SEC=\"512\"", "Line type [part] is not disk: NAME=\"/dev/xvda1\" TYPE=\"part\" SIZE=\"268434390528\" FSTYPE=\"ext4\" LOG-SEC=\"512\"", "filename [xvda1] is a partition", "Disk [/dev/xvda] attrs [{'fstype': '', 'type': 'disk', 'ssize': '512', 'size': '268435456000'}] has partitions" ] } TASK [Set unused_disks if necessary] ******************************************* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:31 Monday 11 November 2024 19:21:31 -0500 (0:00:00.619) 0:00:14.738 ******* ok: [managed-node3] => { "ansible_facts": { "unused_disks": [ "sda", "sdb", "sdc", "sdd", "sde", "sdf", "sdg", "sdh", "sdi", "sdj" ] }, "changed": false } TASK [Print unused disks] ****************************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:36 Monday 11 November 2024 19:21:31 -0500 (0:00:00.061) 0:00:14.800 ******* ok: [managed-node3] => { "unused_disks": [ "sda", "sdb", "sdc", "sdd", "sde", "sdf", "sdg", "sdh", "sdi", "sdj" ] } TASK [Print info from find_unused_disk] **************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:44 Monday 11 November 2024 19:21:32 -0500 (0:00:00.081) 0:00:14.881 ******* skipping: [managed-node3] => {} TASK [Show disk information] *************************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:49 Monday 11 November 2024 19:21:32 -0500 (0:00:00.052) 0:00:14.934 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Exit playbook when there's not enough unused disks in the system] ******** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:58 Monday 11 November 2024 19:21:32 -0500 (0:00:00.051) 0:00:14.986 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Create LVM logical volumes under volume groups] ************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tasks/setup.yml:31 Monday 11 November 2024 19:21:32 -0500 (0:00:00.052) 0:00:15.039 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Monday 11 November 2024 19:21:32 -0500 (0:00:00.121) 0:00:15.160 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Monday 11 November 2024 19:21:32 -0500 (0:00:00.089) 0:00:15.250 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Monday 11 November 2024 19:21:32 -0500 (0:00:00.053) 0:00:15.304 ******* skipping: [managed-node3] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node3] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node3] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed-node3] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Monday 11 November 2024 19:21:32 -0500 (0:00:00.102) 0:00:15.406 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Monday 11 November 2024 19:21:32 -0500 (0:00:00.043) 0:00:15.450 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Monday 11 November 2024 19:21:32 -0500 (0:00:00.063) 0:00:15.513 ******* ok: [managed-node3] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Monday 11 November 2024 19:21:32 -0500 (0:00:00.032) 0:00:15.546 ******* ok: [managed-node3] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Monday 11 November 2024 19:21:32 -0500 (0:00:00.034) 0:00:15.580 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Monday 11 November 2024 19:21:32 -0500 (0:00:00.083) 0:00:15.664 ******* ok: [managed-node3] => { "changed": false, "rc": 0, "results": [ "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed" ] } lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34 TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Monday 11 November 2024 19:21:36 -0500 (0:00:03.904) 0:00:19.569 ******* ok: [managed-node3] => { "storage_pools": [ { "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "name": "test_vg1", "volumes": [ { "name": "thin_pool_dev", "size": "2g", "thin": true, "thin_pool_name": "thin_pool_vg1", "thin_pool_size": "4g" }, { "name": "lv2", "size": "15m", "thin": true, "thin_pool_name": "thin_pool_vg1", "thin_pool_size": "2g" }, { "name": "lv3", "size": "15m", "thin": true, "thin_pool_name": "thin_pool_vg1", "thin_pool_size": "2g" }, { "name": "lv4", "size": "15m", "thin": true, "thin_pool_name": "thin_pool_vg1", "thin_pool_size": "2g" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Monday 11 November 2024 19:21:36 -0500 (0:00:00.043) 0:00:19.612 ******* ok: [managed-node3] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Monday 11 November 2024 19:21:36 -0500 (0:00:00.035) 0:00:19.648 ******* ok: [managed-node3] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Monday 11 November 2024 19:21:40 -0500 (0:00:03.760) 0:00:23.409 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node3 TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Monday 11 November 2024 19:21:40 -0500 (0:00:00.065) 0:00:23.474 ******* TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Monday 11 November 2024 19:21:40 -0500 (0:00:00.032) 0:00:23.507 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************ task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19 Monday 11 November 2024 19:21:40 -0500 (0:00:00.034) 0:00:23.541 ******* TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Monday 11 November 2024 19:21:40 -0500 (0:00:00.032) 0:00:23.573 ******* ok: [managed-node3] => { "changed": false, "rc": 0, "results": [ "7:lvm2-2.02.187-6.el7_9.5.x86_64 providing lvm2 is already installed", "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed" ] } lsrpackages: kpartx lvm2 TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Monday 11 November 2024 19:21:41 -0500 (0:00:00.595) 0:00:24.169 ******* ok: [managed-node3] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "brandbot.service": { "name": "brandbot.service", "source": "systemd", "state": "inactive", "status": "static" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-shell.service": { "name": "console-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.import1.service": { "name": "dbus-org.freedesktop.import1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.machine1.service": { "name": "dbus-org.freedesktop.machine1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "running", "status": "static" }, "dmraid-activation.service": { "name": "dmraid-activation.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-lvmetad.service": { "name": "lvm2-lvmetad.service", "source": "systemd", "state": "running", "status": "static" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "lvm2-pvscan@8:0.service": { "name": "lvm2-pvscan@8:0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:112.service": { "name": "lvm2-pvscan@8:112.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:128.service": { "name": "lvm2-pvscan@8:128.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:144.service": { "name": "lvm2-pvscan@8:144.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:16.service": { "name": "lvm2-pvscan@8:16.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:32.service": { "name": "lvm2-pvscan@8:32.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:48.service": { "name": "lvm2-pvscan@8:48.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:64.service": { "name": "lvm2-pvscan@8:64.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:80.service": { "name": "lvm2-pvscan@8:80.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:96.service": { "name": "lvm2-pvscan@8:96.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "netconsole": { "name": "netconsole", "source": "sysv", "state": "stopped", "status": "disabled" }, "network": { "name": "network", "source": "sysv", "state": "running", "status": "enabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-config.service": { "name": "nfs-config.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-idmap.service": { "name": "nfs-idmap.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-lock.service": { "name": "nfs-lock.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-rquotad.service": { "name": "nfs-rquotad.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-secure.service": { "name": "nfs-secure.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs.service": { "name": "nfs.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfslock.service": { "name": "nfslock.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "postfix.service": { "name": "postfix.service", "source": "systemd", "state": "running", "status": "enabled" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rhel-autorelabel-mark.service": { "name": "rhel-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-autorelabel.service": { "name": "rhel-autorelabel.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-configure.service": { "name": "rhel-configure.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-dmesg.service": { "name": "rhel-dmesg.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-domainname.service": { "name": "rhel-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-import-state.service": { "name": "rhel-import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-loadmodules.service": { "name": "rhel-loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-readonly.service": { "name": "rhel-readonly.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-rquotad.service": { "name": "rpc-rquotad.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpcgssd.service": { "name": "rpcgssd.service", "source": "systemd", "state": "inactive", "status": "static" }, "rpcidmapd.service": { "name": "rpcidmapd.service", "source": "systemd", "state": "inactive", "status": "static" }, "rsyncd.service": { "name": "rsyncd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyncd@.service": { "name": "rsyncd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-policy-migrate-local-changes@.service": { "name": "selinux-policy-migrate-local-changes@.service", "source": "systemd", "state": "unknown", "status": "static" }, "selinux-policy-migrate-local-changes@targeted.service": { "name": "selinux-policy-migrate-local-changes@targeted.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "unknown" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "static" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bootchart.service": { "name": "systemd-bootchart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-importd.service": { "name": "systemd-importd.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machined.service": { "name": "systemd-machined.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-nspawn@.service": { "name": "systemd-nspawn@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-readahead-collect.service": { "name": "systemd-readahead-collect.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-readahead-done.service": { "name": "systemd-readahead-done.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "systemd-readahead-drop.service": { "name": "systemd-readahead-drop.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "systemd-readahead-replay.service": { "name": "systemd-readahead-replay.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill@.service": { "name": "systemd-rfkill@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-shutdownd.service": { "name": "systemd-shutdownd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "wpa_supplicant.service": { "name": "wpa_supplicant.service", "source": "systemd", "state": "inactive", "status": "disabled" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Monday 11 November 2024 19:21:42 -0500 (0:00:00.991) 0:00:25.160 ******* ok: [managed-node3] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Monday 11 November 2024 19:21:42 -0500 (0:00:00.062) 0:00:25.223 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Monday 11 November 2024 19:21:42 -0500 (0:00:00.043) 0:00:25.267 ******* changed: [managed-node3] => { "actions": [ { "action": "create format", "device": "/dev/sde", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sdd", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sdc", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sdb", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/test_vg1", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/test_vg1-thin_pool_vg1", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/test_vg1-lv4", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/test_vg1-lv4", "fs_type": "xfs" }, { "action": "create device", "device": "/dev/mapper/test_vg1-lv3", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/test_vg1-lv3", "fs_type": "xfs" }, { "action": "create device", "device": "/dev/mapper/test_vg1-lv2", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/test_vg1-lv2", "fs_type": "xfs" }, { "action": "create device", "device": "/dev/mapper/test_vg1-thin_pool_dev", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/test_vg1-thin_pool_dev", "fs_type": "xfs" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sdk", "/dev/sdl", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/sdj", "/dev/xvda1", "/dev/mapper/test_vg1-thin_pool_dev", "/dev/mapper/test_vg1-lv2", "/dev/mapper/test_vg1-lv3", "/dev/mapper/test_vg1-lv4" ], "mounts": [], "packages": [ "xfsprogs", "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "test_vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/test_vg1-thin_pool_dev", "_kernel_device": "/dev/dm-7", "_mount_id": "/dev/mapper/test_vg1-thin_pool_dev", "_raw_device": "/dev/mapper/test_vg1-thin_pool_dev", "_raw_kernel_device": "/dev/dm-7", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "mount_user": null, "name": "thin_pool_dev", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "2g", "state": "present", "thin": true, "thin_pool_name": "thin_pool_vg1", "thin_pool_size": "4g", "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv2", "_kernel_device": "/dev/dm-6", "_mount_id": "/dev/mapper/test_vg1-lv2", "_raw_device": "/dev/mapper/test_vg1-lv2", "_raw_kernel_device": "/dev/dm-6", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "mount_user": null, "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "15m", "state": "present", "thin": true, "thin_pool_name": "thin_pool_vg1", "thin_pool_size": "2g", "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv3", "_kernel_device": "/dev/dm-5", "_mount_id": "/dev/mapper/test_vg1-lv3", "_raw_device": "/dev/mapper/test_vg1-lv3", "_raw_kernel_device": "/dev/dm-5", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "mount_user": null, "name": "lv3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "15m", "state": "present", "thin": true, "thin_pool_name": "thin_pool_vg1", "thin_pool_size": "2g", "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv4", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/test_vg1-lv4", "_raw_device": "/dev/mapper/test_vg1-lv4", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "mount_user": null, "name": "lv4", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "15m", "state": "present", "thin": true, "thin_pool_name": "thin_pool_vg1", "thin_pool_size": "2g", "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Monday 11 November 2024 19:21:48 -0500 (0:00:05.834) 0:00:31.101 ******* ok: [managed-node3] => { "changed": false, "cmd": [ "udevadm", "trigger", "--subsystem-match=block" ], "delta": "0:00:00.010509", "end": "2024-11-11 19:21:48.738166", "rc": 0, "start": "2024-11-11 19:21:48.727657" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Monday 11 November 2024 19:21:48 -0500 (0:00:00.561) 0:00:31.663 ******* ok: [managed-node3] => { "changed": false, "stat": { "atime": 1731370735.4243228, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "72884e3f126482c2d28276ff7c57744fa95eff91", "ctime": 1731370722.1001987, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263592, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1731370722.0991988, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1229, "uid": 0, "version": "18446744072116901260", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Monday 11 November 2024 19:21:49 -0500 (0:00:00.302) 0:00:31.965 ******* ok: [managed-node3] => { "backup": "", "changed": false } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Monday 11 November 2024 19:21:49 -0500 (0:00:00.527) 0:00:32.493 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Monday 11 November 2024 19:21:49 -0500 (0:00:00.048) 0:00:32.542 ******* ok: [managed-node3] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/sde", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sdd", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sdc", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sdb", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/test_vg1", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/test_vg1-thin_pool_vg1", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/test_vg1-lv4", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/test_vg1-lv4", "fs_type": "xfs" }, { "action": "create device", "device": "/dev/mapper/test_vg1-lv3", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/test_vg1-lv3", "fs_type": "xfs" }, { "action": "create device", "device": "/dev/mapper/test_vg1-lv2", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/test_vg1-lv2", "fs_type": "xfs" }, { "action": "create device", "device": "/dev/mapper/test_vg1-thin_pool_dev", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/test_vg1-thin_pool_dev", "fs_type": "xfs" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sdk", "/dev/sdl", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/sdj", "/dev/xvda1", "/dev/mapper/test_vg1-thin_pool_dev", "/dev/mapper/test_vg1-lv2", "/dev/mapper/test_vg1-lv3", "/dev/mapper/test_vg1-lv4" ], "mounts": [], "packages": [ "xfsprogs", "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "test_vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/test_vg1-thin_pool_dev", "_kernel_device": "/dev/dm-7", "_mount_id": "/dev/mapper/test_vg1-thin_pool_dev", "_raw_device": "/dev/mapper/test_vg1-thin_pool_dev", "_raw_kernel_device": "/dev/dm-7", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "mount_user": null, "name": "thin_pool_dev", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "2g", "state": "present", "thin": true, "thin_pool_name": "thin_pool_vg1", "thin_pool_size": "4g", "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv2", "_kernel_device": "/dev/dm-6", "_mount_id": "/dev/mapper/test_vg1-lv2", "_raw_device": "/dev/mapper/test_vg1-lv2", "_raw_kernel_device": "/dev/dm-6", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "mount_user": null, "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "15m", "state": "present", "thin": true, "thin_pool_name": "thin_pool_vg1", "thin_pool_size": "2g", "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv3", "_kernel_device": "/dev/dm-5", "_mount_id": "/dev/mapper/test_vg1-lv3", "_raw_device": "/dev/mapper/test_vg1-lv3", "_raw_kernel_device": "/dev/dm-5", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "mount_user": null, "name": "lv3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "15m", "state": "present", "thin": true, "thin_pool_name": "thin_pool_vg1", "thin_pool_size": "2g", "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv4", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/test_vg1-lv4", "_raw_device": "/dev/mapper/test_vg1-lv4", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "mount_user": null, "name": "lv4", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "15m", "state": "present", "thin": true, "thin_pool_name": "thin_pool_vg1", "thin_pool_size": "2g", "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Monday 11 November 2024 19:21:49 -0500 (0:00:00.084) 0:00:32.626 ******* ok: [managed-node3] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "test_vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/test_vg1-thin_pool_dev", "_kernel_device": "/dev/dm-7", "_mount_id": "/dev/mapper/test_vg1-thin_pool_dev", "_raw_device": "/dev/mapper/test_vg1-thin_pool_dev", "_raw_kernel_device": "/dev/dm-7", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "mount_user": null, "name": "thin_pool_dev", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "2g", "state": "present", "thin": true, "thin_pool_name": "thin_pool_vg1", "thin_pool_size": "4g", "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv2", "_kernel_device": "/dev/dm-6", "_mount_id": "/dev/mapper/test_vg1-lv2", "_raw_device": "/dev/mapper/test_vg1-lv2", "_raw_kernel_device": "/dev/dm-6", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "mount_user": null, "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "15m", "state": "present", "thin": true, "thin_pool_name": "thin_pool_vg1", "thin_pool_size": "2g", "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv3", "_kernel_device": "/dev/dm-5", "_mount_id": "/dev/mapper/test_vg1-lv3", "_raw_device": "/dev/mapper/test_vg1-lv3", "_raw_kernel_device": "/dev/dm-5", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "mount_user": null, "name": "lv3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "15m", "state": "present", "thin": true, "thin_pool_name": "thin_pool_vg1", "thin_pool_size": "2g", "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv4", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/test_vg1-lv4", "_raw_device": "/dev/mapper/test_vg1-lv4", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "mount_user": null, "name": "lv4", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "15m", "state": "present", "thin": true, "thin_pool_name": "thin_pool_vg1", "thin_pool_size": "2g", "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Monday 11 November 2024 19:21:49 -0500 (0:00:00.079) 0:00:32.706 ******* ok: [managed-node3] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Monday 11 November 2024 19:21:49 -0500 (0:00:00.056) 0:00:32.762 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Monday 11 November 2024 19:21:49 -0500 (0:00:00.054) 0:00:32.817 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Monday 11 November 2024 19:21:50 -0500 (0:00:00.099) 0:00:32.916 ******* TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Monday 11 November 2024 19:21:50 -0500 (0:00:00.053) 0:00:32.970 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Monday 11 November 2024 19:21:50 -0500 (0:00:00.055) 0:00:33.025 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Monday 11 November 2024 19:21:50 -0500 (0:00:00.053) 0:00:33.078 ******* ok: [managed-node3] => { "changed": false, "stat": { "atime": 1731370641.178454, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1718879272.062, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131079, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1718879026.308, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744072852913879", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Monday 11 November 2024 19:21:50 -0500 (0:00:00.340) 0:00:33.419 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Monday 11 November 2024 19:21:50 -0500 (0:00:00.033) 0:00:33.453 ******* ok: [managed-node3] TASK [Run the snapshot role to create snapshot set of LVs] ********************* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tests_set_thin_basic.yml:50 Monday 11 November 2024 19:21:51 -0500 (0:00:00.769) 0:00:34.222 ******* TASK [fedora.linux_system_roles.snapshot : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:3 Monday 11 November 2024 19:21:51 -0500 (0:00:00.119) 0:00:34.342 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml for managed-node3 TASK [fedora.linux_system_roles.snapshot : Ensure ansible_facts used by role] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:2 Monday 11 November 2024 19:21:51 -0500 (0:00:00.106) 0:00:34.448 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Check if system is ostree] ********** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:10 Monday 11 November 2024 19:21:51 -0500 (0:00:00.071) 0:00:34.520 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Set flag to indicate system is ostree] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:15 Monday 11 November 2024 19:21:51 -0500 (0:00:00.055) 0:00:34.576 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:19 Monday 11 November 2024 19:21:51 -0500 (0:00:00.056) 0:00:34.632 ******* skipping: [managed-node3] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node3] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node3] => (item=CentOS_7.yml) => { "ansible_facts": { "__snapshot_packages": [ "lvm2", "util-linux" ], "__snapshot_python": "/usr/bin/python" }, "ansible_included_var_files": [ "/tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed-node3] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Ensure required packages are installed] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:6 Monday 11 November 2024 19:21:51 -0500 (0:00:00.135) 0:00:34.768 ******* ok: [managed-node3] => { "changed": false, "rc": 0, "results": [ "7:lvm2-2.02.187-6.el7_9.5.x86_64 providing lvm2 is already installed", "util-linux-2.23.2-65.el7_9.1.x86_64 providing util-linux is already installed" ] } lsrpackages: lvm2 util-linux TASK [fedora.linux_system_roles.snapshot : Run snapshot module snapshot] ******* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:14 Monday 11 November 2024 19:21:52 -0500 (0:00:00.693) 0:00:35.462 ******* [WARNING]: The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change. changed: [managed-node3] => { "changed": true, "errors": "", "message": "", "return_code": 0 } TASK [fedora.linux_system_roles.snapshot : Print out response] ***************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:40 Monday 11 November 2024 19:21:53 -0500 (0:00:01.172) 0:00:36.634 ******* ok: [managed-node3] => { "snapshot_cmd": { "changed": true, "errors": "", "failed": false, "message": "", "msg": "", "return_code": 0, "warnings": [ "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change.", "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change." ] } } TASK [fedora.linux_system_roles.snapshot : Set result] ************************* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:45 Monday 11 November 2024 19:21:53 -0500 (0:00:00.076) 0:00:36.711 ******* ok: [managed-node3] => { "ansible_facts": { "snapshot_cmd": { "changed": true, "errors": "", "failed": false, "message": "", "msg": "", "return_code": 0, "warnings": [ "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change.", "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change." ] } }, "changed": false } TASK [fedora.linux_system_roles.snapshot : Set snapshot_facts to the JSON results] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:49 Monday 11 November 2024 19:21:53 -0500 (0:00:00.071) 0:00:36.782 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Show errors] ************************ task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:54 Monday 11 November 2024 19:21:54 -0500 (0:00:00.072) 0:00:36.855 ******* skipping: [managed-node3] => {} TASK [Assert changes for create snapset] *************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tests_set_thin_basic.yml:57 Monday 11 November 2024 19:21:54 -0500 (0:00:00.062) 0:00:36.917 ******* ok: [managed-node3] => { "changed": false } MSG: All assertions passed TASK [Run the snapshot role to verify the set of snapshots for the LVs] ******** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tests_set_thin_basic.yml:61 Monday 11 November 2024 19:21:54 -0500 (0:00:00.061) 0:00:36.979 ******* TASK [fedora.linux_system_roles.snapshot : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:3 Monday 11 November 2024 19:21:54 -0500 (0:00:00.135) 0:00:37.115 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml for managed-node3 TASK [fedora.linux_system_roles.snapshot : Ensure ansible_facts used by role] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:2 Monday 11 November 2024 19:21:54 -0500 (0:00:00.170) 0:00:37.285 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Check if system is ostree] ********** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:10 Monday 11 November 2024 19:21:54 -0500 (0:00:00.075) 0:00:37.361 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Set flag to indicate system is ostree] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:15 Monday 11 November 2024 19:21:54 -0500 (0:00:00.060) 0:00:37.422 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:19 Monday 11 November 2024 19:21:54 -0500 (0:00:00.056) 0:00:37.478 ******* skipping: [managed-node3] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node3] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node3] => (item=CentOS_7.yml) => { "ansible_facts": { "__snapshot_packages": [ "lvm2", "util-linux" ], "__snapshot_python": "/usr/bin/python" }, "ansible_included_var_files": [ "/tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed-node3] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Ensure required packages are installed] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:6 Monday 11 November 2024 19:21:54 -0500 (0:00:00.118) 0:00:37.597 ******* ok: [managed-node3] => { "changed": false, "rc": 0, "results": [ "7:lvm2-2.02.187-6.el7_9.5.x86_64 providing lvm2 is already installed", "util-linux-2.23.2-65.el7_9.1.x86_64 providing util-linux is already installed" ] } lsrpackages: lvm2 util-linux TASK [fedora.linux_system_roles.snapshot : Run snapshot module check] ********** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:14 Monday 11 November 2024 19:21:55 -0500 (0:00:00.737) 0:00:38.334 ******* ok: [managed-node3] => { "changed": false, "errors": "", "message": "", "return_code": 0 } TASK [fedora.linux_system_roles.snapshot : Print out response] ***************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:40 Monday 11 November 2024 19:21:56 -0500 (0:00:00.635) 0:00:38.969 ******* ok: [managed-node3] => { "snapshot_cmd": { "changed": false, "errors": "", "failed": false, "message": "", "msg": "", "return_code": 0, "warnings": [ "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change.", "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change." ] } } TASK [fedora.linux_system_roles.snapshot : Set result] ************************* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:45 Monday 11 November 2024 19:21:56 -0500 (0:00:00.073) 0:00:39.043 ******* ok: [managed-node3] => { "ansible_facts": { "snapshot_cmd": { "changed": false, "errors": "", "failed": false, "message": "", "msg": "", "return_code": 0, "warnings": [ "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change.", "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change." ] } }, "changed": false } TASK [fedora.linux_system_roles.snapshot : Set snapshot_facts to the JSON results] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:49 Monday 11 November 2024 19:21:56 -0500 (0:00:00.083) 0:00:39.127 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Show errors] ************************ task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:54 Monday 11 November 2024 19:21:56 -0500 (0:00:00.066) 0:00:39.194 ******* skipping: [managed-node3] => {} TASK [Create snapset again for idempotence] ************************************ task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tests_set_thin_basic.yml:69 Monday 11 November 2024 19:21:56 -0500 (0:00:00.058) 0:00:39.253 ******* TASK [fedora.linux_system_roles.snapshot : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:3 Monday 11 November 2024 19:21:56 -0500 (0:00:00.130) 0:00:39.383 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml for managed-node3 TASK [fedora.linux_system_roles.snapshot : Ensure ansible_facts used by role] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:2 Monday 11 November 2024 19:21:56 -0500 (0:00:00.098) 0:00:39.482 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Check if system is ostree] ********** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:10 Monday 11 November 2024 19:21:56 -0500 (0:00:00.073) 0:00:39.556 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Set flag to indicate system is ostree] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:15 Monday 11 November 2024 19:21:56 -0500 (0:00:00.058) 0:00:39.614 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:19 Monday 11 November 2024 19:21:56 -0500 (0:00:00.106) 0:00:39.720 ******* skipping: [managed-node3] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node3] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node3] => (item=CentOS_7.yml) => { "ansible_facts": { "__snapshot_packages": [ "lvm2", "util-linux" ], "__snapshot_python": "/usr/bin/python" }, "ansible_included_var_files": [ "/tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed-node3] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Ensure required packages are installed] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:6 Monday 11 November 2024 19:21:57 -0500 (0:00:00.152) 0:00:39.873 ******* ok: [managed-node3] => { "changed": false, "rc": 0, "results": [ "7:lvm2-2.02.187-6.el7_9.5.x86_64 providing lvm2 is already installed", "util-linux-2.23.2-65.el7_9.1.x86_64 providing util-linux is already installed" ] } lsrpackages: lvm2 util-linux TASK [fedora.linux_system_roles.snapshot : Run snapshot module snapshot] ******* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:14 Monday 11 November 2024 19:21:57 -0500 (0:00:00.779) 0:00:40.652 ******* ok: [managed-node3] => { "changed": false, "errors": "snapshot already exists: test_vg1/lv2_snapset1", "message": "", "return_code": 0 } MSG: snapshot already exists: test_vg1/lv2_snapset1 TASK [fedora.linux_system_roles.snapshot : Print out response] ***************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:40 Monday 11 November 2024 19:21:58 -0500 (0:00:00.573) 0:00:41.225 ******* ok: [managed-node3] => { "snapshot_cmd": { "changed": false, "errors": "snapshot already exists: test_vg1/lv2_snapset1", "failed": false, "message": "", "msg": "snapshot already exists: test_vg1/lv2_snapset1", "return_code": 0, "warnings": [ "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change.", "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change." ] } } TASK [fedora.linux_system_roles.snapshot : Set result] ************************* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:45 Monday 11 November 2024 19:21:58 -0500 (0:00:00.063) 0:00:41.288 ******* ok: [managed-node3] => { "ansible_facts": { "snapshot_cmd": { "changed": false, "errors": "snapshot already exists: test_vg1/lv2_snapset1", "failed": false, "message": "", "msg": "snapshot already exists: test_vg1/lv2_snapset1", "return_code": 0, "warnings": [ "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change.", "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change." ] } }, "changed": false } TASK [fedora.linux_system_roles.snapshot : Set snapshot_facts to the JSON results] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:49 Monday 11 November 2024 19:21:58 -0500 (0:00:00.066) 0:00:41.355 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Show errors] ************************ task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:54 Monday 11 November 2024 19:21:58 -0500 (0:00:00.064) 0:00:41.419 ******* skipping: [managed-node3] => {} TASK [Assert no changes for create snapset] ************************************ task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tests_set_thin_basic.yml:76 Monday 11 November 2024 19:21:58 -0500 (0:00:00.059) 0:00:41.479 ******* ok: [managed-node3] => { "changed": false } MSG: All assertions passed TASK [Run the snapshot role remove the set] ************************************ task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tests_set_thin_basic.yml:80 Monday 11 November 2024 19:21:58 -0500 (0:00:00.060) 0:00:41.539 ******* TASK [fedora.linux_system_roles.snapshot : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:3 Monday 11 November 2024 19:21:58 -0500 (0:00:00.140) 0:00:41.679 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml for managed-node3 TASK [fedora.linux_system_roles.snapshot : Ensure ansible_facts used by role] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:2 Monday 11 November 2024 19:21:58 -0500 (0:00:00.075) 0:00:41.755 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Check if system is ostree] ********** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:10 Monday 11 November 2024 19:21:58 -0500 (0:00:00.047) 0:00:41.802 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Set flag to indicate system is ostree] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:15 Monday 11 November 2024 19:21:59 -0500 (0:00:00.057) 0:00:41.860 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:19 Monday 11 November 2024 19:21:59 -0500 (0:00:00.060) 0:00:41.920 ******* skipping: [managed-node3] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node3] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node3] => (item=CentOS_7.yml) => { "ansible_facts": { "__snapshot_packages": [ "lvm2", "util-linux" ], "__snapshot_python": "/usr/bin/python" }, "ansible_included_var_files": [ "/tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed-node3] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Ensure required packages are installed] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:6 Monday 11 November 2024 19:21:59 -0500 (0:00:00.145) 0:00:42.065 ******* ok: [managed-node3] => { "changed": false, "rc": 0, "results": [ "7:lvm2-2.02.187-6.el7_9.5.x86_64 providing lvm2 is already installed", "util-linux-2.23.2-65.el7_9.1.x86_64 providing util-linux is already installed" ] } lsrpackages: lvm2 util-linux TASK [fedora.linux_system_roles.snapshot : Run snapshot module remove] ********* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:14 Monday 11 November 2024 19:21:59 -0500 (0:00:00.748) 0:00:42.814 ******* changed: [managed-node3] => { "changed": true, "errors": "", "message": "", "return_code": 0 } TASK [fedora.linux_system_roles.snapshot : Print out response] ***************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:40 Monday 11 November 2024 19:22:00 -0500 (0:00:00.916) 0:00:43.730 ******* ok: [managed-node3] => { "snapshot_cmd": { "changed": true, "errors": "", "failed": false, "message": "", "msg": "", "return_code": 0, "warnings": [ "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change.", "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change." ] } } TASK [fedora.linux_system_roles.snapshot : Set result] ************************* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:45 Monday 11 November 2024 19:22:00 -0500 (0:00:00.056) 0:00:43.787 ******* ok: [managed-node3] => { "ansible_facts": { "snapshot_cmd": { "changed": true, "errors": "", "failed": false, "message": "", "msg": "", "return_code": 0, "warnings": [ "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change.", "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change." ] } }, "changed": false } TASK [fedora.linux_system_roles.snapshot : Set snapshot_facts to the JSON results] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:49 Monday 11 November 2024 19:22:01 -0500 (0:00:00.060) 0:00:43.847 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Show errors] ************************ task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:54 Monday 11 November 2024 19:22:01 -0500 (0:00:00.077) 0:00:43.924 ******* skipping: [managed-node3] => {} TASK [Assert changes for remove snapset] *************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tests_set_thin_basic.yml:87 Monday 11 November 2024 19:22:01 -0500 (0:00:00.059) 0:00:43.983 ******* ok: [managed-node3] => { "changed": false } MSG: All assertions passed TASK [Run the snapshot role to verify the set is removed] ********************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tests_set_thin_basic.yml:91 Monday 11 November 2024 19:22:01 -0500 (0:00:00.057) 0:00:44.041 ******* TASK [fedora.linux_system_roles.snapshot : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:3 Monday 11 November 2024 19:22:01 -0500 (0:00:00.163) 0:00:44.205 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml for managed-node3 TASK [fedora.linux_system_roles.snapshot : Ensure ansible_facts used by role] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:2 Monday 11 November 2024 19:22:01 -0500 (0:00:00.095) 0:00:44.301 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Check if system is ostree] ********** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:10 Monday 11 November 2024 19:22:01 -0500 (0:00:00.091) 0:00:44.392 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Set flag to indicate system is ostree] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:15 Monday 11 November 2024 19:22:01 -0500 (0:00:00.075) 0:00:44.468 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:19 Monday 11 November 2024 19:22:01 -0500 (0:00:00.073) 0:00:44.542 ******* skipping: [managed-node3] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node3] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node3] => (item=CentOS_7.yml) => { "ansible_facts": { "__snapshot_packages": [ "lvm2", "util-linux" ], "__snapshot_python": "/usr/bin/python" }, "ansible_included_var_files": [ "/tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed-node3] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Ensure required packages are installed] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:6 Monday 11 November 2024 19:22:01 -0500 (0:00:00.156) 0:00:44.698 ******* ok: [managed-node3] => { "changed": false, "rc": 0, "results": [ "7:lvm2-2.02.187-6.el7_9.5.x86_64 providing lvm2 is already installed", "util-linux-2.23.2-65.el7_9.1.x86_64 providing util-linux is already installed" ] } lsrpackages: lvm2 util-linux TASK [fedora.linux_system_roles.snapshot : Run snapshot module remove] ********* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:14 Monday 11 November 2024 19:22:02 -0500 (0:00:00.744) 0:00:45.442 ******* ok: [managed-node3] => { "changed": false, "errors": "", "message": "", "return_code": 0 } TASK [fedora.linux_system_roles.snapshot : Print out response] ***************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:40 Monday 11 November 2024 19:22:03 -0500 (0:00:00.456) 0:00:45.899 ******* ok: [managed-node3] => { "snapshot_cmd": { "changed": false, "errors": "", "failed": false, "message": "", "msg": "", "return_code": 0, "warnings": [ "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change.", "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change." ] } } TASK [fedora.linux_system_roles.snapshot : Set result] ************************* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:45 Monday 11 November 2024 19:22:03 -0500 (0:00:00.051) 0:00:45.950 ******* ok: [managed-node3] => { "ansible_facts": { "snapshot_cmd": { "changed": false, "errors": "", "failed": false, "message": "", "msg": "", "return_code": 0, "warnings": [ "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change.", "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change." ] } }, "changed": false } TASK [fedora.linux_system_roles.snapshot : Set snapshot_facts to the JSON results] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:49 Monday 11 November 2024 19:22:03 -0500 (0:00:00.042) 0:00:45.993 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Show errors] ************************ task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:54 Monday 11 November 2024 19:22:03 -0500 (0:00:00.036) 0:00:46.030 ******* skipping: [managed-node3] => {} TASK [Remove again to check idempotence] *************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tests_set_thin_basic.yml:99 Monday 11 November 2024 19:22:03 -0500 (0:00:00.039) 0:00:46.069 ******* TASK [fedora.linux_system_roles.snapshot : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:3 Monday 11 November 2024 19:22:03 -0500 (0:00:00.099) 0:00:46.169 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml for managed-node3 TASK [fedora.linux_system_roles.snapshot : Ensure ansible_facts used by role] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:2 Monday 11 November 2024 19:22:03 -0500 (0:00:00.059) 0:00:46.229 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Check if system is ostree] ********** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:10 Monday 11 November 2024 19:22:03 -0500 (0:00:00.050) 0:00:46.279 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Set flag to indicate system is ostree] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:15 Monday 11 November 2024 19:22:03 -0500 (0:00:00.037) 0:00:46.317 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/set_vars.yml:19 Monday 11 November 2024 19:22:03 -0500 (0:00:00.035) 0:00:46.352 ******* skipping: [managed-node3] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node3] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node3] => (item=CentOS_7.yml) => { "ansible_facts": { "__snapshot_packages": [ "lvm2", "util-linux" ], "__snapshot_python": "/usr/bin/python" }, "ansible_included_var_files": [ "/tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed-node3] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Ensure required packages are installed] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:6 Monday 11 November 2024 19:22:03 -0500 (0:00:00.086) 0:00:46.439 ******* ok: [managed-node3] => { "changed": false, "rc": 0, "results": [ "7:lvm2-2.02.187-6.el7_9.5.x86_64 providing lvm2 is already installed", "util-linux-2.23.2-65.el7_9.1.x86_64 providing util-linux is already installed" ] } lsrpackages: lvm2 util-linux TASK [fedora.linux_system_roles.snapshot : Run snapshot module remove] ********* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:14 Monday 11 November 2024 19:22:04 -0500 (0:00:00.592) 0:00:47.031 ******* ok: [managed-node3] => { "changed": false, "errors": "", "message": "", "return_code": 0 } TASK [fedora.linux_system_roles.snapshot : Print out response] ***************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:40 Monday 11 November 2024 19:22:04 -0500 (0:00:00.492) 0:00:47.524 ******* ok: [managed-node3] => { "snapshot_cmd": { "changed": false, "errors": "", "failed": false, "message": "", "msg": "", "return_code": 0, "warnings": [ "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change.", "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change." ] } } TASK [fedora.linux_system_roles.snapshot : Set result] ************************* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:45 Monday 11 November 2024 19:22:04 -0500 (0:00:00.041) 0:00:47.566 ******* ok: [managed-node3] => { "ansible_facts": { "snapshot_cmd": { "changed": false, "errors": "", "failed": false, "message": "", "msg": "", "return_code": 0, "warnings": [ "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change.", "The value 20 (type int) in a string field was converted to u'20' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change." ] } }, "changed": false } TASK [fedora.linux_system_roles.snapshot : Set snapshot_facts to the JSON results] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:49 Monday 11 November 2024 19:22:04 -0500 (0:00:00.042) 0:00:47.608 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.snapshot : Show errors] ************************ task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:54 Monday 11 November 2024 19:22:04 -0500 (0:00:00.037) 0:00:47.645 ******* skipping: [managed-node3] => {} TASK [Assert no changes for remove snapset] ************************************ task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tests_set_thin_basic.yml:106 Monday 11 November 2024 19:22:04 -0500 (0:00:00.038) 0:00:47.684 ******* ok: [managed-node3] => { "changed": false } MSG: All assertions passed TASK [Cleanup] ***************************************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tests_set_thin_basic.yml:110 Monday 11 November 2024 19:22:04 -0500 (0:00:00.040) 0:00:47.724 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tasks/cleanup.yml for managed-node3 TASK [Remove storage volumes] ************************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tasks/cleanup.yml:7 Monday 11 November 2024 19:22:04 -0500 (0:00:00.084) 0:00:47.809 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Monday 11 November 2024 19:22:05 -0500 (0:00:00.065) 0:00:47.875 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed-node3 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Monday 11 November 2024 19:22:05 -0500 (0:00:00.060) 0:00:47.935 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Monday 11 November 2024 19:22:05 -0500 (0:00:00.054) 0:00:47.990 ******* skipping: [managed-node3] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node3] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node3] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed-node3] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Monday 11 November 2024 19:22:05 -0500 (0:00:00.098) 0:00:48.089 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Monday 11 November 2024 19:22:05 -0500 (0:00:00.037) 0:00:48.127 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Monday 11 November 2024 19:22:05 -0500 (0:00:00.037) 0:00:48.164 ******* ok: [managed-node3] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Monday 11 November 2024 19:22:05 -0500 (0:00:00.039) 0:00:48.203 ******* ok: [managed-node3] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Monday 11 November 2024 19:22:05 -0500 (0:00:00.039) 0:00:48.243 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed-node3 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Monday 11 November 2024 19:22:05 -0500 (0:00:00.087) 0:00:48.330 ******* ok: [managed-node3] => { "changed": false, "rc": 0, "results": [ "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed" ] } lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34 TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Monday 11 November 2024 19:22:09 -0500 (0:00:03.930) 0:00:52.261 ******* ok: [managed-node3] => { "storage_pools": [ { "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "name": "test_vg1", "state": "absent", "volumes": [ { "name": "thin_pool_dev", "state": "absent" }, { "name": "lv2", "state": "absent" }, { "name": "lv3", "state": "absent" }, { "name": "lv4", "state": "absent" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Monday 11 November 2024 19:22:09 -0500 (0:00:00.077) 0:00:52.339 ******* ok: [managed-node3] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Monday 11 November 2024 19:22:09 -0500 (0:00:00.057) 0:00:52.397 ******* ok: [managed-node3] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Monday 11 November 2024 19:22:14 -0500 (0:00:05.017) 0:00:57.414 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed-node3 TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Monday 11 November 2024 19:22:14 -0500 (0:00:00.138) 0:00:57.553 ******* TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Monday 11 November 2024 19:22:14 -0500 (0:00:00.043) 0:00:57.596 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************ task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19 Monday 11 November 2024 19:22:14 -0500 (0:00:00.039) 0:00:57.635 ******* TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Monday 11 November 2024 19:22:14 -0500 (0:00:00.038) 0:00:57.674 ******* ok: [managed-node3] => { "changed": false, "rc": 0, "results": [ "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed" ] } lsrpackages: kpartx TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Monday 11 November 2024 19:22:15 -0500 (0:00:00.681) 0:00:58.356 ******* ok: [managed-node3] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "brandbot.service": { "name": "brandbot.service", "source": "systemd", "state": "inactive", "status": "static" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-shell.service": { "name": "console-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.import1.service": { "name": "dbus-org.freedesktop.import1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.machine1.service": { "name": "dbus-org.freedesktop.machine1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "running", "status": "static" }, "dmraid-activation.service": { "name": "dmraid-activation.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-lvmetad.service": { "name": "lvm2-lvmetad.service", "source": "systemd", "state": "running", "status": "static" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "lvm2-pvscan@8:0.service": { "name": "lvm2-pvscan@8:0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:112.service": { "name": "lvm2-pvscan@8:112.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:128.service": { "name": "lvm2-pvscan@8:128.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:144.service": { "name": "lvm2-pvscan@8:144.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:16.service": { "name": "lvm2-pvscan@8:16.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:32.service": { "name": "lvm2-pvscan@8:32.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:48.service": { "name": "lvm2-pvscan@8:48.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:64.service": { "name": "lvm2-pvscan@8:64.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:80.service": { "name": "lvm2-pvscan@8:80.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "lvm2-pvscan@8:96.service": { "name": "lvm2-pvscan@8:96.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "netconsole": { "name": "netconsole", "source": "sysv", "state": "stopped", "status": "disabled" }, "network": { "name": "network", "source": "sysv", "state": "running", "status": "enabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-config.service": { "name": "nfs-config.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-idmap.service": { "name": "nfs-idmap.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-lock.service": { "name": "nfs-lock.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-rquotad.service": { "name": "nfs-rquotad.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-secure.service": { "name": "nfs-secure.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs.service": { "name": "nfs.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfslock.service": { "name": "nfslock.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "postfix.service": { "name": "postfix.service", "source": "systemd", "state": "running", "status": "enabled" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rhel-autorelabel-mark.service": { "name": "rhel-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-autorelabel.service": { "name": "rhel-autorelabel.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-configure.service": { "name": "rhel-configure.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-dmesg.service": { "name": "rhel-dmesg.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-domainname.service": { "name": "rhel-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-import-state.service": { "name": "rhel-import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-loadmodules.service": { "name": "rhel-loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-readonly.service": { "name": "rhel-readonly.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-rquotad.service": { "name": "rpc-rquotad.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpcgssd.service": { "name": "rpcgssd.service", "source": "systemd", "state": "inactive", "status": "static" }, "rpcidmapd.service": { "name": "rpcidmapd.service", "source": "systemd", "state": "inactive", "status": "static" }, "rsyncd.service": { "name": "rsyncd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyncd@.service": { "name": "rsyncd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-policy-migrate-local-changes@.service": { "name": "selinux-policy-migrate-local-changes@.service", "source": "systemd", "state": "unknown", "status": "static" }, "selinux-policy-migrate-local-changes@targeted.service": { "name": "selinux-policy-migrate-local-changes@targeted.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "unknown" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "static" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bootchart.service": { "name": "systemd-bootchart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-importd.service": { "name": "systemd-importd.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machined.service": { "name": "systemd-machined.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-nspawn@.service": { "name": "systemd-nspawn@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-readahead-collect.service": { "name": "systemd-readahead-collect.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-readahead-done.service": { "name": "systemd-readahead-done.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "systemd-readahead-drop.service": { "name": "systemd-readahead-drop.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "systemd-readahead-replay.service": { "name": "systemd-readahead-replay.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill@.service": { "name": "systemd-rfkill@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-shutdownd.service": { "name": "systemd-shutdownd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "wpa_supplicant.service": { "name": "wpa_supplicant.service", "source": "systemd", "state": "inactive", "status": "disabled" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Monday 11 November 2024 19:22:16 -0500 (0:00:00.939) 0:00:59.296 ******* ok: [managed-node3] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Monday 11 November 2024 19:22:16 -0500 (0:00:00.078) 0:00:59.374 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Monday 11 November 2024 19:22:16 -0500 (0:00:00.059) 0:00:59.434 ******* changed: [managed-node3] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/test_vg1-lv4", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/test_vg1-lv4", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/test_vg1-lv3", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/test_vg1-lv3", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/test_vg1-lv2", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/test_vg1-lv2", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/test_vg1-thin_pool_dev", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/test_vg1-thin_pool_dev", "fs_type": null }, { "action": "destroy device", "device": "/dev/mapper/test_vg1-thin_pool_vg1", "fs_type": null }, { "action": "destroy device", "device": "/dev/test_vg1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdc", "fs_type": "lvmpv" }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "destroy format", "device": "/dev/sde", "fs_type": "lvmpv" }, { "action": "destroy format", "device": "/dev/sdd", "fs_type": "lvmpv" }, { "action": "destroy format", "device": "/dev/sdb", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdk", "/dev/sdl", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/sdj", "/dev/xvda1" ], "mounts": [], "packages": [ "e2fsprogs" ], "pools": [ { "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "test_vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/test_vg1-thin_pool_dev", "_mount_id": "/dev/mapper/test_vg1-thin_pool_dev", "_raw_device": "/dev/mapper/test_vg1-thin_pool_dev", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "mount_user": null, "name": "thin_pool_dev", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 2147483648, "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv2", "_mount_id": "/dev/mapper/test_vg1-lv2", "_raw_device": "/dev/mapper/test_vg1-lv2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "mount_user": null, "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 16777216, "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv3", "_mount_id": "/dev/mapper/test_vg1-lv3", "_raw_device": "/dev/mapper/test_vg1-lv3", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "mount_user": null, "name": "lv3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 16777216, "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv4", "_mount_id": "/dev/mapper/test_vg1-lv4", "_raw_device": "/dev/mapper/test_vg1-lv4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "mount_user": null, "name": "lv4", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 16777216, "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Monday 11 November 2024 19:22:24 -0500 (0:00:08.029) 0:01:07.463 ******* ok: [managed-node3] => { "changed": false, "cmd": [ "udevadm", "trigger", "--subsystem-match=block" ], "delta": "0:00:00.009542", "end": "2024-11-11 19:22:24.935547", "rc": 0, "start": "2024-11-11 19:22:24.926005" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Monday 11 November 2024 19:22:25 -0500 (0:00:00.396) 0:01:07.859 ******* ok: [managed-node3] => { "changed": false, "stat": { "atime": 1731370735.4243228, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "72884e3f126482c2d28276ff7c57744fa95eff91", "ctime": 1731370722.1001987, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263592, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1731370722.0991988, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1229, "uid": 0, "version": "18446744072116901260", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Monday 11 November 2024 19:22:25 -0500 (0:00:00.362) 0:01:08.222 ******* ok: [managed-node3] => { "backup": "", "changed": false } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Monday 11 November 2024 19:22:25 -0500 (0:00:00.348) 0:01:08.570 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Monday 11 November 2024 19:22:25 -0500 (0:00:00.056) 0:01:08.627 ******* ok: [managed-node3] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/test_vg1-lv4", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/test_vg1-lv4", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/test_vg1-lv3", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/test_vg1-lv3", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/test_vg1-lv2", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/test_vg1-lv2", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/test_vg1-thin_pool_dev", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/test_vg1-thin_pool_dev", "fs_type": null }, { "action": "destroy device", "device": "/dev/mapper/test_vg1-thin_pool_vg1", "fs_type": null }, { "action": "destroy device", "device": "/dev/test_vg1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdc", "fs_type": "lvmpv" }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "destroy format", "device": "/dev/sde", "fs_type": "lvmpv" }, { "action": "destroy format", "device": "/dev/sdd", "fs_type": "lvmpv" }, { "action": "destroy format", "device": "/dev/sdb", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdk", "/dev/sdl", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/sdj", "/dev/xvda1" ], "mounts": [], "packages": [ "e2fsprogs" ], "pools": [ { "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "test_vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/test_vg1-thin_pool_dev", "_mount_id": "/dev/mapper/test_vg1-thin_pool_dev", "_raw_device": "/dev/mapper/test_vg1-thin_pool_dev", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "mount_user": null, "name": "thin_pool_dev", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 2147483648, "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv2", "_mount_id": "/dev/mapper/test_vg1-lv2", "_raw_device": "/dev/mapper/test_vg1-lv2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "mount_user": null, "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 16777216, "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv3", "_mount_id": "/dev/mapper/test_vg1-lv3", "_raw_device": "/dev/mapper/test_vg1-lv3", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "mount_user": null, "name": "lv3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 16777216, "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv4", "_mount_id": "/dev/mapper/test_vg1-lv4", "_raw_device": "/dev/mapper/test_vg1-lv4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "mount_user": null, "name": "lv4", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 16777216, "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Monday 11 November 2024 19:22:25 -0500 (0:00:00.111) 0:01:08.739 ******* ok: [managed-node3] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "test_vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/test_vg1-thin_pool_dev", "_mount_id": "/dev/mapper/test_vg1-thin_pool_dev", "_raw_device": "/dev/mapper/test_vg1-thin_pool_dev", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "mount_user": null, "name": "thin_pool_dev", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 2147483648, "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv2", "_mount_id": "/dev/mapper/test_vg1-lv2", "_raw_device": "/dev/mapper/test_vg1-lv2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "mount_user": null, "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 16777216, "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv3", "_mount_id": "/dev/mapper/test_vg1-lv3", "_raw_device": "/dev/mapper/test_vg1-lv3", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "mount_user": null, "name": "lv3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 16777216, "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/test_vg1-lv4", "_mount_id": "/dev/mapper/test_vg1-lv4", "_raw_device": "/dev/mapper/test_vg1-lv4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc", "sdd", "sde" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "mount_user": null, "name": "lv4", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 16777216, "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Monday 11 November 2024 19:22:26 -0500 (0:00:00.087) 0:01:08.826 ******* ok: [managed-node3] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Monday 11 November 2024 19:22:26 -0500 (0:00:00.053) 0:01:08.880 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Monday 11 November 2024 19:22:26 -0500 (0:00:00.045) 0:01:08.926 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Monday 11 November 2024 19:22:26 -0500 (0:00:00.046) 0:01:08.973 ******* TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Monday 11 November 2024 19:22:26 -0500 (0:00:00.039) 0:01:09.012 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Monday 11 November 2024 19:22:26 -0500 (0:00:00.077) 0:01:09.089 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Monday 11 November 2024 19:22:26 -0500 (0:00:00.041) 0:01:09.131 ******* ok: [managed-node3] => { "changed": false, "stat": { "atime": 1731370641.178454, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1718879272.062, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131079, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1718879026.308, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744072852913879", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Monday 11 November 2024 19:22:26 -0500 (0:00:00.344) 0:01:09.476 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Monday 11 November 2024 19:22:26 -0500 (0:00:00.061) 0:01:09.538 ******* ok: [managed-node3] TASK [Save unused_disk_return before verify] *********************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tasks/cleanup.yml:30 Monday 11 November 2024 19:22:27 -0500 (0:00:00.667) 0:01:10.205 ******* ok: [managed-node3] => { "ansible_facts": { "unused_disks_before": [ "sda", "sdb", "sdc", "sdd", "sde", "sdf", "sdg", "sdh", "sdi", "sdj" ] }, "changed": false } TASK [Verify that pools/volumes used in test are removed] ********************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tasks/cleanup.yml:34 Monday 11 November 2024 19:22:27 -0500 (0:00:00.040) 0:01:10.245 ******* included: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml for managed-node3 TASK [Check if system is ostree] *********************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:5 Monday 11 November 2024 19:22:27 -0500 (0:00:00.085) 0:01:10.330 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set flag to indicate system is ostree] *********************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:10 Monday 11 November 2024 19:22:27 -0500 (0:00:00.050) 0:01:10.381 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure test packages] **************************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:14 Monday 11 November 2024 19:22:27 -0500 (0:00:00.056) 0:01:10.437 ******* ok: [managed-node3] => { "changed": false, "rc": 0, "results": [ "util-linux-2.23.2-65.el7_9.1.x86_64 providing util-linux is already installed" ] } lsrpackages: util-linux TASK [Find unused disks in the system] ***************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:23 Monday 11 November 2024 19:22:28 -0500 (0:00:00.673) 0:01:11.110 ******* ok: [managed-node3] => { "changed": false, "disks": [ "sda", "sdb", "sdc", "sdd", "sde", "sdf", "sdg", "sdh", "sdi", "sdj" ], "info": [ "Line: NAME=\"/dev/sda\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdb\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdc\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdd\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sde\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdf\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdg\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdh\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdi\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdj\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdk\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdl\" TYPE=\"disk\" SIZE=\"1073741824\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/xvda\" TYPE=\"disk\" SIZE=\"268435456000\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/xvda1\" TYPE=\"part\" SIZE=\"268434390528\" FSTYPE=\"ext4\" LOG-SEC=\"512\"", "Line type [part] is not disk: NAME=\"/dev/xvda1\" TYPE=\"part\" SIZE=\"268434390528\" FSTYPE=\"ext4\" LOG-SEC=\"512\"", "filename [xvda1] is a partition", "Disk [/dev/xvda] attrs [{'fstype': '', 'type': 'disk', 'ssize': '512', 'size': '268435456000'}] has partitions" ] } TASK [Set unused_disks if necessary] ******************************************* task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:31 Monday 11 November 2024 19:22:28 -0500 (0:00:00.373) 0:01:11.484 ******* ok: [managed-node3] => { "ansible_facts": { "unused_disks": [ "sda", "sdb", "sdc", "sdd", "sde", "sdf", "sdg", "sdh", "sdi", "sdj" ] }, "changed": false } TASK [Print unused disks] ****************************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:36 Monday 11 November 2024 19:22:28 -0500 (0:00:00.070) 0:01:11.554 ******* ok: [managed-node3] => { "unused_disks": [ "sda", "sdb", "sdc", "sdd", "sde", "sdf", "sdg", "sdh", "sdi", "sdj" ] } TASK [Print info from find_unused_disk] **************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:44 Monday 11 November 2024 19:22:28 -0500 (0:00:00.061) 0:01:11.616 ******* skipping: [managed-node3] => {} TASK [Show disk information] *************************************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:49 Monday 11 November 2024 19:22:28 -0500 (0:00:00.101) 0:01:11.717 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Exit playbook when there's not enough unused disks in the system] ******** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:58 Monday 11 November 2024 19:22:28 -0500 (0:00:00.087) 0:01:11.805 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Debug why list of unused disks has changed] ****************************** task path: /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tasks/cleanup.yml:40 Monday 11 November 2024 19:22:29 -0500 (0:00:00.061) 0:01:11.866 ******* skipping: [managed-node3] => { "changed": false, "skip_reason": "Conditional result was False" } META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* managed-node3 : ok=124 changed=4 unreachable=0 failed=0 skipped=81 rescued=0 ignored=0 Monday 11 November 2024 19:22:29 -0500 (0:00:00.084) 0:01:11.951 ******* =============================================================================== fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 8.03s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 5.83s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Get required packages --------------- 5.02s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 fedora.linux_system_roles.storage : Make sure blivet is available ------- 4.13s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 fedora.linux_system_roles.storage : Make sure blivet is available ------- 3.93s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 fedora.linux_system_roles.storage : Make sure blivet is available ------- 3.90s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 fedora.linux_system_roles.storage : Get required packages --------------- 3.76s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 fedora.linux_system_roles.storage : Get service facts ------------------- 1.18s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Gathering Facts --------------------------------------------------------- 1.05s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/tests_set_thin_basic.yml:2 fedora.linux_system_roles.storage : Get service facts ------------------- 0.99s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 fedora.linux_system_roles.storage : Get service facts ------------------- 0.94s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 fedora.linux_system_roles.storage : Update facts ------------------------ 0.91s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 fedora.linux_system_roles.storage : Make sure required packages are installed --- 0.84s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 fedora.linux_system_roles.snapshot : Ensure required packages are installed --- 0.78s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:6 fedora.linux_system_roles.storage : Update facts ------------------------ 0.77s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 fedora.linux_system_roles.storage : Get required packages --------------- 0.75s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 fedora.linux_system_roles.snapshot : Ensure required packages are installed --- 0.74s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:6 fedora.linux_system_roles.snapshot : Ensure required packages are installed --- 0.74s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/snapshot/tasks/main.yml:6 fedora.linux_system_roles.storage : Make sure required packages are installed --- 0.68s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Ensure test packages ---------------------------------------------------- 0.67s /tmp/collections-IzJ/ansible_collections/fedora/linux_system_roles/tests/snapshot/get_unused_disk.yml:14