ansible-playbook [core 2.12.6] config file = /etc/ansible/ansible.cfg configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.9/site-packages/ansible ansible collection location = /tmp/tmp5bkr4li_ executable location = /usr/bin/ansible-playbook python version = 3.9.13 (main, May 18 2022, 00:00:00) [GCC 11.3.1 20220421 (Red Hat 11.3.1-2)] jinja version = 2.11.3 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: rhel-7_setup.yml ***************************************************** 1 plays in /cache/rhel-7_setup.yml PLAY [Setup repos] ************************************************************* META: ran handlers TASK [set up internal repositories] ******************************************** task path: /cache/rhel-7_setup.yml:5 Thursday 21 July 2022 14:58:38 +0000 (0:00:00.018) 0:00:00.018 ********* changed: [/cache/rhel-7.qcow2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [/cache/rhel-7.qcow2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [/cache/rhel-7.qcow2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [/cache/rhel-7.qcow2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [/cache/rhel-7.qcow2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* /cache/rhel-7.qcow2 : ok=1 changed=1 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 Thursday 21 July 2022 14:58:39 +0000 (0:00:01.423) 0:00:01.441 ********* =============================================================================== set up internal repositories -------------------------------------------- 1.42s /cache/rhel-7_setup.yml:5 ----------------------------------------------------- statically imported: /tmp/tmpaxjje44y/tests/create-test-file.yml statically imported: /tmp/tmpaxjje44y/tests/verify-data-preservation.yml statically imported: /tmp/tmpaxjje44y/tests/create-test-file.yml statically imported: /tmp/tmpaxjje44y/tests/verify-data-preservation.yml statically imported: /tmp/tmpaxjje44y/tests/create-test-file.yml statically imported: /tmp/tmpaxjje44y/tests/verify-data-preservation.yml PLAYBOOK: tests_luks_pool_nvme_generated.yml *********************************** 2 plays in /tmp/tmpaxjje44y/tests/tests_luks_pool_nvme_generated.yml PLAY [all] ********************************************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/tmpaxjje44y/tests/tests_luks_pool_nvme_generated.yml:3 Thursday 21 July 2022 14:58:39 +0000 (0:00:00.029) 0:00:01.471 ********* ok: [/cache/rhel-7.qcow2] META: ran handlers TASK [set disk interface for test] ********************************************* task path: /tmp/tmpaxjje44y/tests/tests_luks_pool_nvme_generated.yml:7 Thursday 21 July 2022 14:58:40 +0000 (0:00:01.049) 0:00:02.520 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_use_interface": "nvme" }, "changed": false } META: ran handlers META: ran handlers PLAY [all] ********************************************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:2 Thursday 21 July 2022 14:58:40 +0000 (0:00:00.052) 0:00:02.573 ********* ok: [/cache/rhel-7.qcow2] META: ran handlers TASK [include_role : linux-system-roles.storage] ******************************* task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:14 Thursday 21 July 2022 14:58:41 +0000 (0:00:00.727) 0:00:03.300 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 14:58:41 +0000 (0:00:00.033) 0:00:03.334 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 14:58:41 +0000 (0:00:00.029) 0:00:03.363 ********* ok: [/cache/rhel-7.qcow2] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 14:58:41 +0000 (0:00:00.401) 0:00:03.765 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 14:58:41 +0000 (0:00:00.055) 0:00:03.820 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 14:58:42 +0000 (0:00:00.029) 0:00:03.850 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 14:58:42 +0000 (0:00:00.028) 0:00:03.878 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 14:58:42 +0000 (0:00:00.046) 0:00:03.924 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 14:58:42 +0000 (0:00:00.019) 0:00:03.944 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "changes": { "installed": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "rc": 0, "results": [ "Loaded plugins: search-disabled-repos\nResolving Dependencies\n--> Running transaction check\n---> Package libblockdev-crypto.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libblockdev-utils(x86-64) = 2.18-5.el7 for package: libblockdev-crypto-2.18-5.el7.x86_64\n--> Processing Dependency: libvolume_key.so.1()(64bit) for package: libblockdev-crypto-2.18-5.el7.x86_64\n--> Processing Dependency: libbd_utils.so.2()(64bit) for package: libblockdev-crypto-2.18-5.el7.x86_64\n---> Package libblockdev-dm.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libdmraid.so.1(Base)(64bit) for package: libblockdev-dm-2.18-5.el7.x86_64\n--> Processing Dependency: dmraid for package: libblockdev-dm-2.18-5.el7.x86_64\n--> Processing Dependency: libdmraid.so.1()(64bit) for package: libblockdev-dm-2.18-5.el7.x86_64\n---> Package libblockdev-lvm.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: lvm2 for package: libblockdev-lvm-2.18-5.el7.x86_64\n--> Processing Dependency: device-mapper-persistent-data for package: libblockdev-lvm-2.18-5.el7.x86_64\n---> Package libblockdev-mdraid.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: mdadm for package: libblockdev-mdraid-2.18-5.el7.x86_64\n--> Processing Dependency: libbytesize.so.1()(64bit) for package: libblockdev-mdraid-2.18-5.el7.x86_64\n---> Package libblockdev-swap.x86_64 0:2.18-5.el7 will be installed\n---> Package python-enum34.noarch 0:1.0.4-1.el7 will be installed\n---> Package python2-blivet3.noarch 1:3.1.3-3.el7 will be installed\n--> Processing Dependency: blivet3-data = 1:3.1.3-3.el7 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-bytesize >= 0.3 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-blockdev >= 2.17 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: pyparted >= 3.9 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-hawkey for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: lsof for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Running transaction check\n---> Package blivet3-data.noarch 1:3.1.3-3.el7 will be installed\n---> Package device-mapper-persistent-data.x86_64 0:0.8.5-3.el7_9.2 will be installed\n--> Processing Dependency: libaio.so.1(LIBAIO_0.4)(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n--> Processing Dependency: libaio.so.1(LIBAIO_0.1)(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n--> Processing Dependency: libaio.so.1()(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n---> Package dmraid.x86_64 0:1.0.0.rc16-28.el7 will be installed\n--> Processing Dependency: libdevmapper-event.so.1.02(Base)(64bit) for package: dmraid-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: dmraid-events for package: dmraid-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: libdevmapper-event.so.1.02()(64bit) for package: dmraid-1.0.0.rc16-28.el7.x86_64\n---> Package libblockdev-utils.x86_64 0:2.18-5.el7 will be installed\n---> Package libbytesize.x86_64 0:1.2-1.el7 will be installed\n--> Processing Dependency: libmpfr.so.4()(64bit) for package: libbytesize-1.2-1.el7.x86_64\n---> Package lsof.x86_64 0:4.87-6.el7 will be installed\n---> Package lvm2.x86_64 7:2.02.187-6.el7_9.5 will be installed\n--> Processing Dependency: lvm2-libs = 7:2.02.187-6.el7_9.5 for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n--> Processing Dependency: liblvm2app.so.2.2(Base)(64bit) for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n--> Processing Dependency: liblvm2app.so.2.2()(64bit) for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n---> Package mdadm.x86_64 0:4.1-9.el7_9 will be installed\n--> Processing Dependency: libreport-filesystem for package: mdadm-4.1-9.el7_9.x86_64\n---> Package pyparted.x86_64 1:3.9-15.el7 will be installed\n---> Package python2-blockdev.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libblockdev(x86-64) = 2.18-5.el7 for package: python2-blockdev-2.18-5.el7.x86_64\n---> Package python2-bytesize.x86_64 0:1.2-1.el7 will be installed\n---> Package python2-hawkey.x86_64 0:0.22.5-2.el7_9 will be installed\n--> Processing Dependency: libdnf(x86-64) = 0.22.5-2.el7_9 for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: python2-libdnf = 0.22.5-2.el7_9 for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolv.so.0(SOLV_1.0)(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolvext.so.0(SOLV_1.0)(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libdnf.so.2()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libjson-glib-1.0.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libmodulemd.so.1()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: librepo.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: librhsm.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolv.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolvext.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n---> Package volume_key-libs.x86_64 0:0.3.9-9.el7 will be installed\n--> Running transaction check\n---> Package device-mapper-event-libs.x86_64 7:1.02.170-6.el7_9.5 will be installed\n---> Package dmraid-events.x86_64 0:1.0.0.rc16-28.el7 will be installed\n--> Processing Dependency: sgpio for package: dmraid-events-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: device-mapper-event for package: dmraid-events-1.0.0.rc16-28.el7.x86_64\n---> Package json-glib.x86_64 0:1.4.2-2.el7 will be installed\n---> Package libaio.x86_64 0:0.3.109-13.el7 will be installed\n---> Package libblockdev.x86_64 0:2.18-5.el7 will be installed\n---> Package libdnf.x86_64 0:0.22.5-2.el7_9 will be installed\n---> Package libmodulemd.x86_64 0:1.6.3-1.el7 will be installed\n---> Package librepo.x86_64 0:1.8.1-8.el7_9 will be installed\n---> Package libreport-filesystem.x86_64 0:2.1.11-53.el7 will be installed\n---> Package librhsm.x86_64 0:0.0.3-3.el7_9 will be installed\n---> Package libsolv.x86_64 0:0.6.34-4.el7 will be installed\n---> Package lvm2-libs.x86_64 7:2.02.187-6.el7_9.5 will be installed\n---> Package mpfr.x86_64 0:3.1.1-4.el7 will be installed\n---> Package python2-libdnf.x86_64 0:0.22.5-2.el7_9 will be installed\n--> Running transaction check\n---> Package device-mapper-event.x86_64 7:1.02.170-6.el7_9.5 will be installed\n---> Package sgpio.x86_64 0:1.2.0.10-13.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package Arch Version Repository Size\n================================================================================\nInstalling:\n libblockdev-crypto x86_64 2.18-5.el7 rhel 60 k\n libblockdev-dm x86_64 2.18-5.el7 rhel-optional 54 k\n libblockdev-lvm x86_64 2.18-5.el7 rhel 62 k\n libblockdev-mdraid x86_64 2.18-5.el7 rhel 57 k\n libblockdev-swap x86_64 2.18-5.el7 rhel 52 k\n python-enum34 noarch 1.0.4-1.el7 rhel 52 k\n python2-blivet3 noarch 1:3.1.3-3.el7 rhel 851 k\nInstalling for dependencies:\n blivet3-data noarch 1:3.1.3-3.el7 rhel 77 k\n device-mapper-event\n x86_64 7:1.02.170-6.el7_9.5 rhel 192 k\n device-mapper-event-libs\n x86_64 7:1.02.170-6.el7_9.5 rhel 192 k\n device-mapper-persistent-data\n x86_64 0.8.5-3.el7_9.2 rhel 423 k\n dmraid x86_64 1.0.0.rc16-28.el7 rhel 151 k\n dmraid-events x86_64 1.0.0.rc16-28.el7 rhel 21 k\n json-glib x86_64 1.4.2-2.el7 rhel 134 k\n libaio x86_64 0.3.109-13.el7 rhel 24 k\n libblockdev x86_64 2.18-5.el7 rhel 119 k\n libblockdev-utils x86_64 2.18-5.el7 rhel 59 k\n libbytesize x86_64 1.2-1.el7 rhel 52 k\n libdnf x86_64 0.22.5-2.el7_9 rhel-7-server-extras-rpms 536 k\n libmodulemd x86_64 1.6.3-1.el7 rhel-7-server-extras-rpms 153 k\n librepo x86_64 1.8.1-8.el7_9 rhel 82 k\n libreport-filesystem\n x86_64 2.1.11-53.el7 rhel 41 k\n librhsm x86_64 0.0.3-3.el7_9 rhel-7-server-extras-rpms 28 k\n libsolv x86_64 0.6.34-4.el7 rhel 329 k\n lsof x86_64 4.87-6.el7 rhel 331 k\n lvm2 x86_64 7:2.02.187-6.el7_9.5 rhel 1.3 M\n lvm2-libs x86_64 7:2.02.187-6.el7_9.5 rhel 1.1 M\n mdadm x86_64 4.1-9.el7_9 rhel 440 k\n mpfr x86_64 3.1.1-4.el7 rhel 203 k\n pyparted x86_64 1:3.9-15.el7 rhel 195 k\n python2-blockdev x86_64 2.18-5.el7 rhel 61 k\n python2-bytesize x86_64 1.2-1.el7 rhel 22 k\n python2-hawkey x86_64 0.22.5-2.el7_9 rhel-7-server-extras-rpms 71 k\n python2-libdnf x86_64 0.22.5-2.el7_9 rhel-7-server-extras-rpms 611 k\n sgpio x86_64 1.2.0.10-13.el7 rhel 14 k\n volume_key-libs x86_64 0.3.9-9.el7 rhel 141 k\n\nTransaction Summary\n================================================================================\nInstall 7 Packages (+29 Dependent packages)\n\nTotal download size: 8.2 M\nInstalled size: 24 M\nDownloading packages:\n--------------------------------------------------------------------------------\nTotal 19 MB/s | 8.2 MB 00:00 \nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n Installing : libblockdev-utils-2.18-5.el7.x86_64 1/36 \n Installing : 7:device-mapper-event-libs-1.02.170-6.el7_9.5.x86_64 2/36 \n Installing : json-glib-1.4.2-2.el7.x86_64 3/36 \n Installing : librhsm-0.0.3-3.el7_9.x86_64 4/36 \n Installing : libsolv-0.6.34-4.el7.x86_64 5/36 \n Installing : libaio-0.3.109-13.el7.x86_64 6/36 \n Installing : librepo-1.8.1-8.el7_9.x86_64 7/36 \n Installing : libmodulemd-1.6.3-1.el7.x86_64 8/36 \n Installing : libdnf-0.22.5-2.el7_9.x86_64 9/36 \n Installing : device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64 10/36 \n Installing : 7:device-mapper-event-1.02.170-6.el7_9.5.x86_64 11/36 \n Installing : 7:lvm2-libs-2.02.187-6.el7_9.5.x86_64 12/36 \n Installing : 7:lvm2-2.02.187-6.el7_9.5.x86_64 13/36 \n Installing : python2-libdnf-0.22.5-2.el7_9.x86_64 14/36 \n Installing : python2-hawkey-0.22.5-2.el7_9.x86_64 15/36 \n Installing : libblockdev-2.18-5.el7.x86_64 16/36 \n Installing : python2-blockdev-2.18-5.el7.x86_64 17/36 \n Installing : 1:pyparted-3.9-15.el7.x86_64 18/36 \n Installing : sgpio-1.2.0.10-13.el7.x86_64 19/36 \n Installing : dmraid-1.0.0.rc16-28.el7.x86_64 20/36 \n Installing : dmraid-events-1.0.0.rc16-28.el7.x86_64 21/36 \n Installing : volume_key-libs-0.3.9-9.el7.x86_64 22/36 \n Installing : mpfr-3.1.1-4.el7.x86_64 23/36 \n Installing : libbytesize-1.2-1.el7.x86_64 24/36 \n Installing : python2-bytesize-1.2-1.el7.x86_64 25/36 \n Installing : libreport-filesystem-2.1.11-53.el7.x86_64 26/36 \n Installing : mdadm-4.1-9.el7_9.x86_64 27/36 \n Installing : 1:blivet3-data-3.1.3-3.el7.noarch 28/36 \n Installing : lsof-4.87-6.el7.x86_64 29/36 \n Installing : 1:python2-blivet3-3.1.3-3.el7.noarch 30/36 \n Installing : libblockdev-mdraid-2.18-5.el7.x86_64 31/36 \n Installing : libblockdev-crypto-2.18-5.el7.x86_64 32/36 \n Installing : libblockdev-dm-2.18-5.el7.x86_64 33/36 \n Installing : libblockdev-lvm-2.18-5.el7.x86_64 34/36 \n Installing : libblockdev-swap-2.18-5.el7.x86_64 35/36 \n Installing : python-enum34-1.0.4-1.el7.noarch 36/36 \n Verifying : 7:device-mapper-event-1.02.170-6.el7_9.5.x86_64 1/36 \n Verifying : libblockdev-swap-2.18-5.el7.x86_64 2/36 \n Verifying : librhsm-0.0.3-3.el7_9.x86_64 3/36 \n Verifying : libblockdev-lvm-2.18-5.el7.x86_64 4/36 \n Verifying : lsof-4.87-6.el7.x86_64 5/36 \n Verifying : libblockdev-mdraid-2.18-5.el7.x86_64 6/36 \n Verifying : libdnf-0.22.5-2.el7_9.x86_64 7/36 \n Verifying : python-enum34-1.0.4-1.el7.noarch 8/36 \n Verifying : 1:blivet3-data-3.1.3-3.el7.noarch 9/36 \n Verifying : dmraid-events-1.0.0.rc16-28.el7.x86_64 10/36 \n Verifying : python2-blockdev-2.18-5.el7.x86_64 11/36 \n Verifying : libmodulemd-1.6.3-1.el7.x86_64 12/36 \n Verifying : librepo-1.8.1-8.el7_9.x86_64 13/36 \n Verifying : libblockdev-dm-2.18-5.el7.x86_64 14/36 \n Verifying : json-glib-1.4.2-2.el7.x86_64 15/36 \n Verifying : libaio-0.3.109-13.el7.x86_64 16/36 \n Verifying : 7:lvm2-libs-2.02.187-6.el7_9.5.x86_64 17/36 \n Verifying : python2-hawkey-0.22.5-2.el7_9.x86_64 18/36 \n Verifying : python2-bytesize-1.2-1.el7.x86_64 19/36 \n Verifying : libblockdev-2.18-5.el7.x86_64 20/36 \n Verifying : libreport-filesystem-2.1.11-53.el7.x86_64 21/36 \n Verifying : libbytesize-1.2-1.el7.x86_64 22/36 \n Verifying : 7:device-mapper-event-libs-1.02.170-6.el7_9.5.x86_64 23/36 \n Verifying : python2-libdnf-0.22.5-2.el7_9.x86_64 24/36 \n Verifying : 7:lvm2-2.02.187-6.el7_9.5.x86_64 25/36 \n Verifying : libblockdev-utils-2.18-5.el7.x86_64 26/36 \n Verifying : mpfr-3.1.1-4.el7.x86_64 27/36 \n Verifying : volume_key-libs-0.3.9-9.el7.x86_64 28/36 \n Verifying : libsolv-0.6.34-4.el7.x86_64 29/36 \n Verifying : device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64 30/36 \n Verifying : 1:python2-blivet3-3.1.3-3.el7.noarch 31/36 \n Verifying : dmraid-1.0.0.rc16-28.el7.x86_64 32/36 \n Verifying : mdadm-4.1-9.el7_9.x86_64 33/36 \n Verifying : sgpio-1.2.0.10-13.el7.x86_64 34/36 \n Verifying : libblockdev-crypto-2.18-5.el7.x86_64 35/36 \n Verifying : 1:pyparted-3.9-15.el7.x86_64 36/36 \n\nInstalled:\n libblockdev-crypto.x86_64 0:2.18-5.el7 libblockdev-dm.x86_64 0:2.18-5.el7 \n libblockdev-lvm.x86_64 0:2.18-5.el7 libblockdev-mdraid.x86_64 0:2.18-5.el7\n libblockdev-swap.x86_64 0:2.18-5.el7 python-enum34.noarch 0:1.0.4-1.el7 \n python2-blivet3.noarch 1:3.1.3-3.el7 \n\nDependency Installed:\n blivet3-data.noarch 1:3.1.3-3.el7 \n device-mapper-event.x86_64 7:1.02.170-6.el7_9.5 \n device-mapper-event-libs.x86_64 7:1.02.170-6.el7_9.5 \n device-mapper-persistent-data.x86_64 0:0.8.5-3.el7_9.2 \n dmraid.x86_64 0:1.0.0.rc16-28.el7 \n dmraid-events.x86_64 0:1.0.0.rc16-28.el7 \n json-glib.x86_64 0:1.4.2-2.el7 \n libaio.x86_64 0:0.3.109-13.el7 \n libblockdev.x86_64 0:2.18-5.el7 \n libblockdev-utils.x86_64 0:2.18-5.el7 \n libbytesize.x86_64 0:1.2-1.el7 \n libdnf.x86_64 0:0.22.5-2.el7_9 \n libmodulemd.x86_64 0:1.6.3-1.el7 \n librepo.x86_64 0:1.8.1-8.el7_9 \n libreport-filesystem.x86_64 0:2.1.11-53.el7 \n librhsm.x86_64 0:0.0.3-3.el7_9 \n libsolv.x86_64 0:0.6.34-4.el7 \n lsof.x86_64 0:4.87-6.el7 \n lvm2.x86_64 7:2.02.187-6.el7_9.5 \n lvm2-libs.x86_64 7:2.02.187-6.el7_9.5 \n mdadm.x86_64 0:4.1-9.el7_9 \n mpfr.x86_64 0:3.1.1-4.el7 \n pyparted.x86_64 1:3.9-15.el7 \n python2-blockdev.x86_64 0:2.18-5.el7 \n python2-bytesize.x86_64 0:1.2-1.el7 \n python2-hawkey.x86_64 0:0.22.5-2.el7_9 \n python2-libdnf.x86_64 0:0.22.5-2.el7_9 \n sgpio.x86_64 0:1.2.0.10-13.el7 \n volume_key-libs.x86_64 0:0.3.9-9.el7 \n\nComplete!\n" ] } TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 14:58:50 +0000 (0:00:08.571) 0:00:12.516 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 14:58:50 +0000 (0:00:00.035) 0:00:12.551 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 14:58:50 +0000 (0:00:00.033) 0:00:12.585 ********* ok: [/cache/rhel-7.qcow2] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 14:58:51 +0000 (0:00:00.621) 0:00:13.206 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 14:58:51 +0000 (0:00:00.043) 0:00:13.250 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 14:58:51 +0000 (0:00:00.031) 0:00:13.281 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 14:58:51 +0000 (0:00:00.037) 0:00:13.318 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 14:58:51 +0000 (0:00:00.032) 0:00:13.351 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [] } TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 14:58:52 +0000 (0:00:00.522) 0:00:13.874 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "brandbot.service": { "name": "brandbot.service", "source": "systemd", "state": "stopped", "status": "static" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-shell.service": { "name": "console-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.import1.service": { "name": "dbus-org.freedesktop.import1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.machine1.service": { "name": "dbus-org.freedesktop.machine1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dmraid-activation.service": { "name": "dmraid-activation.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "exim.service": { "name": "exim.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "ip6tables.service": { "name": "ip6tables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iptables.service": { "name": "iptables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmetad.service": { "name": "lvm2-lvmetad.service", "source": "systemd", "state": "running", "status": "static" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "netconsole": { "name": "netconsole", "source": "sysv", "state": "stopped", "status": "disabled" }, "network": { "name": "network", "source": "sysv", "state": "running", "status": "enabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "active" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-config.service": { "name": "nfs-config.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-idmap.service": { "name": "nfs-idmap.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-lock.service": { "name": "nfs-lock.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-rquotad.service": { "name": "nfs-rquotad.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-secure-server.service": { "name": "nfs-secure-server.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-secure.service": { "name": "nfs-secure.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs.service": { "name": "nfs.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfslock.service": { "name": "nfslock.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ovirt-guest-agent.service": { "name": "ovirt-guest-agent.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "postfix.service": { "name": "postfix.service", "source": "systemd", "state": "running", "status": "enabled" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhel-autorelabel-mark.service": { "name": "rhel-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-autorelabel.service": { "name": "rhel-autorelabel.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-configure.service": { "name": "rhel-configure.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-dmesg.service": { "name": "rhel-dmesg.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-domainname.service": { "name": "rhel-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-import-state.service": { "name": "rhel-import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-loadmodules.service": { "name": "rhel-loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-readonly.service": { "name": "rhel-readonly.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhnsd": { "name": "rhnsd", "source": "sysv", "state": "running", "status": "enabled" }, "rhnsd.service": { "name": "rhnsd.service", "source": "systemd", "state": "running", "status": "active" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-rquotad.service": { "name": "rpc-rquotad.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpcgssd.service": { "name": "rpcgssd.service", "source": "systemd", "state": "inactive", "status": "static" }, "rpcidmapd.service": { "name": "rpcidmapd.service", "source": "systemd", "state": "inactive", "status": "static" }, "rsyncd.service": { "name": "rsyncd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyncd@.service": { "name": "rsyncd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-policy-migrate-local-changes@.service": { "name": "selinux-policy-migrate-local-changes@.service", "source": "systemd", "state": "unknown", "status": "static" }, "selinux-policy-migrate-local-changes@targeted.service": { "name": "selinux-policy-migrate-local-changes@targeted.service", "source": "systemd", "state": "stopped", "status": "active" }, "sendmail.service": { "name": "sendmail.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "static" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bootchart.service": { "name": "systemd-bootchart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-importd.service": { "name": "systemd-importd.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machined.service": { "name": "systemd-machined.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-nspawn@.service": { "name": "systemd-nspawn@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-readahead-collect.service": { "name": "systemd-readahead-collect.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-readahead-done.service": { "name": "systemd-readahead-done.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "systemd-readahead-drop.service": { "name": "systemd-readahead-drop.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "systemd-readahead-replay.service": { "name": "systemd-readahead-replay.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill@.service": { "name": "systemd-rfkill@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-shutdownd.service": { "name": "systemd-shutdownd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "wpa_supplicant.service": { "name": "wpa_supplicant.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 14:58:53 +0000 (0:00:01.137) 0:00:15.012 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 14:58:53 +0000 (0:00:00.057) 0:00:15.069 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 14:58:53 +0000 (0:00:00.020) 0:00:15.089 ********* ok: [/cache/rhel-7.qcow2] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 14:58:53 +0000 (0:00:00.467) 0:00:15.557 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 14:58:53 +0000 (0:00:00.037) 0:00:15.594 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 14:58:53 +0000 (0:00:00.022) 0:00:15.617 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 14:58:53 +0000 (0:00:00.039) 0:00:15.656 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 14:58:53 +0000 (0:00:00.035) 0:00:15.692 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 14:58:53 +0000 (0:00:00.036) 0:00:15.728 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 14:58:53 +0000 (0:00:00.035) 0:00:15.764 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 14:58:53 +0000 (0:00:00.025) 0:00:15.789 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 14:58:54 +0000 (0:00:00.033) 0:00:15.823 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 14:58:54 +0000 (0:00:00.024) 0:00:15.847 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658415528.2181506, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658201031.524, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 70, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658200515.884, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744071677828413", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 14:58:54 +0000 (0:00:00.434) 0:00:16.282 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 14:58:54 +0000 (0:00:00.023) 0:00:16.305 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [Mark tasks to be skipped] ************************************************ task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:17 Thursday 21 July 2022 14:58:55 +0000 (0:00:00.869) 0:00:17.174 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_skip_checks": [ "blivet_available", "service_facts" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:23 Thursday 21 July 2022 14:58:55 +0000 (0:00:00.068) 0:00:17.243 ********* included: /tmp/tmpaxjje44y/tests/get_unused_disk.yml for /cache/rhel-7.qcow2 TASK [Find unused disks in the system] ***************************************** task path: /tmp/tmpaxjje44y/tests/get_unused_disk.yml:2 Thursday 21 July 2022 14:58:55 +0000 (0:00:00.061) 0:00:17.305 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "disks": [ "nvme1n1" ] } TASK [Set unused_disks if necessary] ******************************************* task path: /tmp/tmpaxjje44y/tests/get_unused_disk.yml:9 Thursday 21 July 2022 14:58:55 +0000 (0:00:00.436) 0:00:17.741 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "unused_disks": [ "nvme1n1" ] }, "changed": false } TASK [Exit playbook when there's not enough unused disks in the system] ******** task path: /tmp/tmpaxjje44y/tests/get_unused_disk.yml:14 Thursday 21 July 2022 14:58:55 +0000 (0:00:00.036) 0:00:17.778 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Print unused disks] ****************************************************** task path: /tmp/tmpaxjje44y/tests/get_unused_disk.yml:19 Thursday 21 July 2022 14:58:55 +0000 (0:00:00.037) 0:00:17.816 ********* ok: [/cache/rhel-7.qcow2] => { "unused_disks": [ "nvme1n1" ] } TASK [Create an encrypted lvm pool] ******************************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:34 Thursday 21 July 2022 14:58:56 +0000 (0:00:00.035) 0:00:17.852 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 14:58:56 +0000 (0:00:00.038) 0:00:17.890 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 14:58:56 +0000 (0:00:00.033) 0:00:17.924 ********* ok: [/cache/rhel-7.qcow2] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 14:58:56 +0000 (0:00:00.415) 0:00:18.340 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 14:58:56 +0000 (0:00:00.061) 0:00:18.401 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 14:58:56 +0000 (0:00:00.033) 0:00:18.434 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 14:58:56 +0000 (0:00:00.031) 0:00:18.466 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 14:58:56 +0000 (0:00:00.044) 0:00:18.510 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 14:58:56 +0000 (0:00:00.021) 0:00:18.531 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 14:58:56 +0000 (0:00:00.027) 0:00:18.559 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "name": "foo", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 14:58:56 +0000 (0:00:00.036) 0:00:18.596 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 14:58:56 +0000 (0:00:00.035) 0:00:18.632 ********* ok: [/cache/rhel-7.qcow2] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "cryptsetup", "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 14:58:57 +0000 (0:00:00.897) 0:00:19.529 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 14:58:57 +0000 (0:00:00.051) 0:00:19.581 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 14:58:57 +0000 (0:00:00.034) 0:00:19.615 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 14:58:57 +0000 (0:00:00.039) 0:00:19.654 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 14:58:57 +0000 (0:00:00.037) 0:00:19.692 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "changes": { "installed": [ "cryptsetup" ] }, "rc": 0, "results": [ "7:lvm2-2.02.187-6.el7_9.5.x86_64 providing lvm2 is already installed", "Loaded plugins: search-disabled-repos\nResolving Dependencies\n--> Running transaction check\n---> Package cryptsetup.x86_64 0:2.0.3-6.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package Arch Version Repository Size\n================================================================================\nInstalling:\n cryptsetup x86_64 2.0.3-6.el7 rhel 154 k\n\nTransaction Summary\n================================================================================\nInstall 1 Package\n\nTotal download size: 154 k\nInstalled size: 354 k\nDownloading packages:\nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n Installing : cryptsetup-2.0.3-6.el7.x86_64 1/1 \n Verifying : cryptsetup-2.0.3-6.el7.x86_64 1/1 \n\nInstalled:\n cryptsetup.x86_64 0:2.0.3-6.el7 \n\nComplete!\n" ] } TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 14:58:59 +0000 (0:00:01.527) 0:00:21.220 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 14:58:59 +0000 (0:00:00.030) 0:00:21.251 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 14:58:59 +0000 (0:00:00.103) 0:00:21.354 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 14:58:59 +0000 (0:00:00.023) 0:00:21.378 ********* fatal: [/cache/rhel-7.qcow2]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: encrypted pool 'foo' missing key/password TASK [linux-system-roles.storage : failed message] ***************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 21 July 2022 14:59:00 +0000 (0:00:00.866) 0:00:22.244 ********* fatal: [/cache/rhel-7.qcow2]: FAILED! => { "changed": false } MSG: {'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': True, 'pools': [{'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': True, 'encryption_key_size': None, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'fs_type': None, 'mount_options': None, 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': None, 'raid_level': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'deduplication': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'raid_disks': [], 'cache_mode': None, 'name': 'test1', 'cached': None, 'type': None, 'thin': False, 'cache_size': None, 'cache_devices': [], 'fs_create_options': None}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}], 'volumes': [], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': [], 'msg': "encrypted pool 'foo' missing key/password", '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 14:59:00 +0000 (0:00:00.039) 0:00:22.284 ********* TASK [Check that we failed in the role] **************************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:53 Thursday 21 July 2022 14:59:00 +0000 (0:00:00.023) 0:00:22.308 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the output of the keyless luks test] ****************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:59 Thursday 21 July 2022 14:59:00 +0000 (0:00:00.033) 0:00:22.341 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Mark tasks to be skipped] ************************************************ task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:66 Thursday 21 July 2022 14:59:00 +0000 (0:00:00.045) 0:00:22.387 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_skip_checks": [ "blivet_available", "packages_installed", "service_facts" ] }, "changed": false } TASK [Create a key file] ******************************************************* task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:75 Thursday 21 July 2022 14:59:00 +0000 (0:00:00.034) 0:00:22.422 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "gid": 0, "group": "root", "mode": "0600", "owner": "root", "path": "/tmp/storage_testxPKADilukskey", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 0, "state": "file", "uid": 0 } TASK [Write the key into the key file] ***************************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:82 Thursday 21 July 2022 14:59:01 +0000 (0:00:00.411) 0:00:22.833 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "checksum": "7a4dff3752e2baf5617c57eaac048e2b95e8af91", "dest": "/tmp/storage_testxPKADilukskey", "gid": 0, "group": "root", "md5sum": "4ac07b967150835c00d0865161e48744", "mode": "0600", "owner": "root", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 32, "src": "/root/.ansible/tmp/ansible-tmp-1658415541.0601985-199645-196034051006107/source", "state": "file", "uid": 0 } TASK [Create an encrypted lvm pool using a key file] *************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:89 Thursday 21 July 2022 14:59:01 +0000 (0:00:00.699) 0:00:23.533 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 14:59:01 +0000 (0:00:00.039) 0:00:23.572 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 14:59:01 +0000 (0:00:00.034) 0:00:23.607 ********* ok: [/cache/rhel-7.qcow2] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 14:59:02 +0000 (0:00:00.417) 0:00:24.024 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 14:59:02 +0000 (0:00:00.094) 0:00:24.119 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 14:59:02 +0000 (0:00:00.036) 0:00:24.155 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 14:59:02 +0000 (0:00:00.034) 0:00:24.190 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 14:59:02 +0000 (0:00:00.044) 0:00:24.234 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 14:59:02 +0000 (0:00:00.023) 0:00:24.258 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 14:59:02 +0000 (0:00:00.027) 0:00:24.285 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_key": "/tmp/storage_testxPKADilukskey", "name": "foo", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 14:59:02 +0000 (0:00:00.068) 0:00:24.354 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 14:59:02 +0000 (0:00:00.069) 0:00:24.423 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 14:59:02 +0000 (0:00:00.030) 0:00:24.454 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 14:59:02 +0000 (0:00:00.028) 0:00:24.482 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 14:59:02 +0000 (0:00:00.035) 0:00:24.517 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 14:59:02 +0000 (0:00:00.033) 0:00:24.551 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 14:59:02 +0000 (0:00:00.089) 0:00:24.641 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 14:59:02 +0000 (0:00:00.022) 0:00:24.663 ********* changed: [/cache/rhel-7.qcow2] => { "actions": [ { "action": "create format", "device": "/dev/nvme1n1", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/nvme1n1", "name": "luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20", "password": "/tmp/storage_testxPKADilukskey", "state": "present" } ], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "cryptsetup", "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": "/tmp/storage_testxPKADilukskey", "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 14:59:10 +0000 (0:00:07.592) 0:00:32.255 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 14:59:10 +0000 (0:00:00.038) 0:00:32.294 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 14:59:10 +0000 (0:00:00.022) 0:00:32.316 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/nvme1n1", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/nvme1n1", "name": "luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20", "password": "/tmp/storage_testxPKADilukskey", "state": "present" } ], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "cryptsetup", "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": "/tmp/storage_testxPKADilukskey", "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 14:59:10 +0000 (0:00:00.037) 0:00:32.354 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": "/tmp/storage_testxPKADilukskey", "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 14:59:10 +0000 (0:00:00.036) 0:00:32.391 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 14:59:10 +0000 (0:00:00.036) 0:00:32.427 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 14:59:10 +0000 (0:00:00.037) 0:00:32.465 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 14:59:11 +0000 (0:00:00.743) 0:00:33.209 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 14:59:11 +0000 (0:00:00.502) 0:00:33.712 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 14:59:12 +0000 (0:00:00.452) 0:00:34.165 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658415528.2181506, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658201031.524, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 70, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658200515.884, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744071677828413", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 14:59:12 +0000 (0:00:00.307) 0:00:34.472 ********* changed: [/cache/rhel-7.qcow2] => (item={'state': 'present', 'password': '/tmp/storage_testxPKADilukskey', 'name': 'luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20', 'backing_device': '/dev/nvme1n1'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/nvme1n1", "name": "luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20", "password": "/tmp/storage_testxPKADilukskey", "state": "present" } } MSG: line added TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 14:59:13 +0000 (0:00:00.444) 0:00:34.916 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:104 Thursday 21 July 2022 14:59:13 +0000 (0:00:00.833) 0:00:35.750 ********* included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1 Thursday 21 July 2022 14:59:14 +0000 (0:00:00.072) 0:00:35.822 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_pools_list": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": "/tmp/storage_testxPKADilukskey", "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6 Thursday 21 July 2022 14:59:14 +0000 (0:00:00.052) 0:00:35.874 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14 Thursday 21 July 2022 14:59:14 +0000 (0:00:00.036) 0:00:35.910 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "4G", "type": "lvm", "uuid": "27f62460-cfe3-42cd-b33c-fbffda6864f8" }, "/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20": { "fstype": "LVM2_member", "label": "", "name": "/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20", "size": "10G", "type": "crypt", "uuid": "5j9HzG-wfHz-pN39-nbqk-rxDn-qlrk-OjaFyc" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "crypto_LUKS", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "a0dd669f-b9b3-4844-8d7f-f75db79f0d20" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-14-58-25-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19 Thursday 21 July 2022 14:59:14 +0000 (0:00:00.399) 0:00:36.310 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003345", "end": "2022-07-21 10:59:14.546121", "rc": 0, "start": "2022-07-21 10:59:14.542776" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24 Thursday 21 July 2022 14:59:14 +0000 (0:00:00.407) 0:00:36.718 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003557", "end": "2022-07-21 10:59:14.858965", "failed_when_result": false, "rc": 0, "start": "2022-07-21 10:59:14.855408" } STDOUT: luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20 /dev/nvme1n1 /tmp/storage_testxPKADilukskey TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33 Thursday 21 July 2022 14:59:15 +0000 (0:00:00.312) 0:00:37.030 ********* included: /tmp/tmpaxjje44y/tests/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': True, 'encryption_key_size': None, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': '/tmp/storage_testxPKADilukskey', 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:5 Thursday 21 July 2022 14:59:15 +0000 (0:00:00.060) 0:00:37.090 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:18 Thursday 21 July 2022 14:59:15 +0000 (0:00:00.037) 0:00:37.128 ********* included: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members) included: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 14:59:15 +0000 (0:00:00.048) 0:00:37.176 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 14:59:15 +0000 (0:00:00.057) 0:00:37.233 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20", "pv": "/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 14:59:15 +0000 (0:00:00.435) 0:00:37.669 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 14:59:15 +0000 (0:00:00.055) 0:00:37.725 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 14:59:15 +0000 (0:00:00.055) 0:00:37.780 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.054) 0:00:37.835 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "crypt" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.043) 0:00:37.878 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.027) 0:00:37.905 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.026) 0:00:37.932 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.086) 0:00:38.018 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-md.yml for /cache/rhel-7.qcow2 TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:6 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.071) 0:00:38.090 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:12 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.023) 0:00:38.114 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:16 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.024) 0:00:38.138 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:20 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.026) 0:00:38.164 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:24 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.025) 0:00:38.189 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:30 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.025) 0:00:38.215 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:36 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.024) 0:00:38.239 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:44 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.024) 0:00:38.264 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.031) 0:00:38.296 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.044) 0:00:38.340 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.044) 0:00:38.385 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.029) 0:00:38.414 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.027) 0:00:38.441 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.029) 0:00:38.471 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml for /cache/rhel-7.qcow2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.044) 0:00:38.515 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.043) 0:00:38.559 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.022) 0:00:38.582 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.021) 0:00:38.603 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.024) 0:00:38.628 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.021) 0:00:38.649 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2 TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.043) 0:00:38.693 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "1", "_storage_test_expected_crypttab_key_file": "/tmp/storage_testxPKADilukskey" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.049) 0:00:38.742 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml for /cache/rhel-7.qcow2 => (item=/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20) TASK [Get the backing device path] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:1 Thursday 21 July 2022 14:59:16 +0000 (0:00:00.045) 0:00:38.787 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "cmd": [ "realpath", "/dev/disk/by-uuid/a0dd669f-b9b3-4844-8d7f-f75db79f0d20" ], "delta": "0:00:00.005491", "end": "2022-07-21 10:59:16.940485", "rc": 0, "start": "2022-07-21 10:59:16.934994" } STDOUT: /dev/nvme1n1 TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:6 Thursday 21 July 2022 14:59:17 +0000 (0:00:00.327) 0:00:39.115 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this member] *************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:11 Thursday 21 July 2022 14:59:17 +0000 (0:00:00.509) 0:00:39.624 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cryptsetup", "luksDump", "/dev/nvme1n1" ], "delta": "0:00:00.099302", "end": "2022-07-21 10:59:17.893788", "rc": 0, "start": "2022-07-21 10:59:17.794486" } STDOUT: LUKS header information for /dev/nvme1n1 Version: 1 Cipher name: aes Cipher mode: xts-plain64 Hash spec: sha256 Payload offset: 4096 MK bits: 512 MK digest: d5 b1 49 1a 1e e6 fd e7 7d 07 4c a4 12 33 f3 6c ae 2a 42 4f MK salt: de 8f 15 70 e6 6f b8 11 36 50 9b ee 87 bc 36 e7 c3 b4 31 c8 35 21 df 80 0e 47 ac b3 a1 55 1f b6 MK iterations: 22413 UUID: a0dd669f-b9b3-4844-8d7f-f75db79f0d20 Key Slot 0: ENABLED Iterations: 359592 Salt: 35 ff 60 1f 3a 0d 88 c5 a9 93 51 01 6d 5a 83 8e 57 46 d6 54 8b 5a 15 67 c1 bb 93 5f ba 82 f6 8a Key material offset: 8 AF stripes: 4000 Key Slot 1: DISABLED Key Slot 2: DISABLED Key Slot 3: DISABLED Key Slot 4: DISABLED Key Slot 5: DISABLED Key Slot 6: DISABLED Key Slot 7: DISABLED TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:17 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.442) 0:00:40.067 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:23 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.026) 0:00:40.094 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:29 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.024) 0:00:40.118 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.022) 0:00:40.141 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.048) 0:00:40.190 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [ "luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20 /dev/nvme1n1 /tmp/storage_testxPKADilukskey" ] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.052) 0:00:40.243 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.049) 0:00:40.292 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.050) 0:00:40.343 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.037) 0:00:40.381 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.055) 0:00:40.436 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.036) 0:00:40.473 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.038) 0:00:40.511 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.051) 0:00:40.562 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.047) 0:00:40.609 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.025) 0:00:40.635 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.024) 0:00:40.659 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.031) 0:00:40.691 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.028) 0:00:40.719 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.025) 0:00:40.745 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.025) 0:00:40.771 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 14:59:18 +0000 (0:00:00.024) 0:00:40.796 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.034) 0:00:40.831 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.033) 0:00:40.864 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.041) 0:00:40.906 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.046) 0:00:40.952 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.079) 0:00:41.032 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.037) 0:00:41.069 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1037768, "block_size": 4096, "block_total": 1046016, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4250697728, "size_total": 4284481536, "uuid": "27f62460-cfe3-42cd-b33c-fbffda6864f8" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1037768, "block_size": 4096, "block_total": 1046016, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4250697728, "size_total": 4284481536, "uuid": "27f62460-cfe3-42cd-b33c-fbffda6864f8" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.088) 0:00:41.158 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.128) 0:00:41.287 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.052) 0:00:41.339 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.052) 0:00:41.392 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.026) 0:00:41.418 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.025) 0:00:41.444 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.024) 0:00:41.469 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.037) 0:00:41.506 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.063) 0:00:41.570 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.050) 0:00:41.620 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.050) 0:00:41.670 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.037) 0:00:41.708 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.035) 0:00:41.744 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.038) 0:00:41.782 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 14:59:19 +0000 (0:00:00.039) 0:00:41.821 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658415550.0471506, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658415550.0471506, "dev": 5, "device_type": 64513, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 33597, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658415550.0471506, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 14:59:20 +0000 (0:00:00.314) 0:00:42.136 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 14:59:20 +0000 (0:00:00.038) 0:00:42.174 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 14:59:20 +0000 (0:00:00.040) 0:00:42.215 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 14:59:20 +0000 (0:00:00.036) 0:00:42.251 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 14:59:20 +0000 (0:00:00.022) 0:00:42.274 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 14:59:20 +0000 (0:00:00.039) 0:00:42.314 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 14:59:20 +0000 (0:00:00.025) 0:00:42.339 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.519) 0:00:42.858 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.027) 0:00:42.885 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.023) 0:00:42.909 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.053) 0:00:42.962 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.026) 0:00:42.988 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.023) 0:00:43.012 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.023) 0:00:43.035 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.023) 0:00:43.059 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.026) 0:00:43.085 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.087) 0:00:43.173 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.051) 0:00:43.224 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.038) 0:00:43.263 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.038) 0:00:43.302 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.038) 0:00:43.340 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.071) 0:00:43.412 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.038) 0:00:43.450 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.036) 0:00:43.487 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.044) 0:00:43.531 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.110) 0:00:43.642 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.040) 0:00:43.682 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.036) 0:00:43.718 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 14:59:21 +0000 (0:00:00.036) 0:00:43.755 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 14:59:22 +0000 (0:00:00.396) 0:00:44.151 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 14:59:22 +0000 (0:00:00.318) 0:00:44.470 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "4294967296" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 14:59:22 +0000 (0:00:00.052) 0:00:44.522 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 14:59:22 +0000 (0:00:00.036) 0:00:44.558 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 14:59:22 +0000 (0:00:00.039) 0:00:44.598 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 14:59:22 +0000 (0:00:00.036) 0:00:44.634 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 14:59:22 +0000 (0:00:00.034) 0:00:44.669 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 14:59:22 +0000 (0:00:00.037) 0:00:44.706 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 14:59:22 +0000 (0:00:00.033) 0:00:44.739 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 14:59:22 +0000 (0:00:00.034) 0:00:44.774 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 14:59:22 +0000 (0:00:00.036) 0:00:44.810 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 14:59:23 +0000 (0:00:00.053) 0:00:44.864 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.023120", "end": "2022-07-21 10:59:23.039925", "rc": 0, "start": "2022-07-21 10:59:23.016805" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 14:59:23 +0000 (0:00:00.351) 0:00:45.216 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 14:59:23 +0000 (0:00:00.051) 0:00:45.267 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 14:59:23 +0000 (0:00:00.052) 0:00:45.320 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 14:59:23 +0000 (0:00:00.036) 0:00:45.356 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 14:59:23 +0000 (0:00:00.037) 0:00:45.393 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 14:59:23 +0000 (0:00:00.038) 0:00:45.432 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 14:59:23 +0000 (0:00:00.039) 0:00:45.472 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43 Thursday 21 July 2022 14:59:23 +0000 (0:00:00.036) 0:00:45.509 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53 Thursday 21 July 2022 14:59:23 +0000 (0:00:00.022) 0:00:45.532 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Remove the key file] ***************************************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:106 Thursday 21 July 2022 14:59:23 +0000 (0:00:00.036) 0:00:45.568 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "path": "/tmp/storage_testxPKADilukskey", "state": "absent" } TASK [create a file] *********************************************************** task path: /tmp/tmpaxjje44y/tests/create-test-file.yml:10 Thursday 21 July 2022 14:59:24 +0000 (0:00:00.452) 0:00:46.021 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "dest": "/opt/test1/quux", "gid": 0, "group": "root", "mode": "0644", "owner": "root", "secontext": "unconfined_u:object_r:unlabeled_t:s0", "size": 0, "state": "file", "uid": 0 } TASK [Remove the encryption layer] ********************************************* task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:116 Thursday 21 July 2022 14:59:24 +0000 (0:00:00.331) 0:00:46.352 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 14:59:24 +0000 (0:00:00.040) 0:00:46.393 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 14:59:24 +0000 (0:00:00.034) 0:00:46.427 ********* ok: [/cache/rhel-7.qcow2] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 14:59:25 +0000 (0:00:00.414) 0:00:46.842 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 14:59:25 +0000 (0:00:00.100) 0:00:46.942 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 14:59:25 +0000 (0:00:00.103) 0:00:47.046 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 14:59:25 +0000 (0:00:00.034) 0:00:47.080 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 14:59:25 +0000 (0:00:00.047) 0:00:47.127 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 14:59:25 +0000 (0:00:00.022) 0:00:47.150 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 14:59:25 +0000 (0:00:00.029) 0:00:47.179 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "nvme1n1" ], "encryption": false, "encryption_password": "yabbadabbadoo", "name": "foo", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 14:59:25 +0000 (0:00:00.039) 0:00:47.218 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 14:59:25 +0000 (0:00:00.042) 0:00:47.260 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 14:59:25 +0000 (0:00:00.030) 0:00:47.291 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 14:59:25 +0000 (0:00:00.029) 0:00:47.321 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 14:59:25 +0000 (0:00:00.027) 0:00:47.348 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 14:59:25 +0000 (0:00:00.029) 0:00:47.378 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 14:59:25 +0000 (0:00:00.058) 0:00:47.436 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 14:59:25 +0000 (0:00:00.022) 0:00:47.458 ********* fatal: [/cache/rhel-7.qcow2]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: cannot remove and recreate existing pool 'foo' in safe mode TASK [linux-system-roles.storage : failed message] ***************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 21 July 2022 14:59:26 +0000 (0:00:01.125) 0:00:48.583 ********* fatal: [/cache/rhel-7.qcow2]: FAILED! => { "changed": false } MSG: {'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': True, 'pools': [{'name': 'foo', 'encryption_password': 'yabbadabbadoo', 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': 0, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': 'luks1', 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'fs_type': None, 'mount_options': None, 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': None, 'raid_level': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'deduplication': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'raid_disks': [], 'cache_mode': None, 'name': 'test1', 'cached': None, 'type': None, 'thin': False, 'cache_size': None, 'cache_devices': [], 'fs_create_options': None}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}], 'volumes': [], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': [], 'msg': "cannot remove and recreate existing pool 'foo' in safe mode", '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 14:59:26 +0000 (0:00:00.044) 0:00:48.628 ********* TASK [Check that we failed in the role] **************************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:134 Thursday 21 July 2022 14:59:26 +0000 (0:00:00.021) 0:00:48.650 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the output of the safe_mode test] ********************************* task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:140 Thursday 21 July 2022 14:59:26 +0000 (0:00:00.038) 0:00:48.688 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [stat the file] *********************************************************** task path: /tmp/tmpaxjje44y/tests/verify-data-preservation.yml:10 Thursday 21 July 2022 14:59:26 +0000 (0:00:00.050) 0:00:48.738 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658415564.1761506, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658415564.1761506, "dev": 64513, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 67, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0644", "mtime": 1658415564.1761506, "nlink": 1, "path": "/opt/test1/quux", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "version": "512842912", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [assert file presence] **************************************************** task path: /tmp/tmpaxjje44y/tests/verify-data-preservation.yml:15 Thursday 21 July 2022 14:59:27 +0000 (0:00:00.326) 0:00:49.065 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Remove the encryption layer] ********************************************* task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:151 Thursday 21 July 2022 14:59:27 +0000 (0:00:00.045) 0:00:49.111 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 14:59:27 +0000 (0:00:00.040) 0:00:49.151 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 14:59:27 +0000 (0:00:00.035) 0:00:49.186 ********* ok: [/cache/rhel-7.qcow2] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 14:59:27 +0000 (0:00:00.417) 0:00:49.604 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 14:59:27 +0000 (0:00:00.110) 0:00:49.714 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 14:59:27 +0000 (0:00:00.075) 0:00:49.790 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 14:59:28 +0000 (0:00:00.035) 0:00:49.826 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 14:59:28 +0000 (0:00:00.049) 0:00:49.875 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 14:59:28 +0000 (0:00:00.023) 0:00:49.898 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 14:59:28 +0000 (0:00:00.030) 0:00:49.929 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "nvme1n1" ], "encryption": false, "encryption_password": "yabbadabbadoo", "name": "foo", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 14:59:28 +0000 (0:00:00.040) 0:00:49.969 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 14:59:28 +0000 (0:00:00.038) 0:00:50.008 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 14:59:28 +0000 (0:00:00.031) 0:00:50.039 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 14:59:28 +0000 (0:00:00.029) 0:00:50.069 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 14:59:28 +0000 (0:00:00.029) 0:00:50.098 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 14:59:28 +0000 (0:00:00.032) 0:00:50.131 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 14:59:28 +0000 (0:00:00.056) 0:00:50.187 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 14:59:28 +0000 (0:00:00.022) 0:00:50.209 ********* changed: [/cache/rhel-7.qcow2] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20", "fs_type": null }, { "action": "destroy format", "device": "/dev/nvme1n1", "fs_type": "luks" }, { "action": "create format", "device": "/dev/nvme1n1", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/nvme1n1", "name": "luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20", "password": "-", "state": "absent" } ], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "nvme1n1" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 14:59:30 +0000 (0:00:02.116) 0:00:52.326 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 14:59:30 +0000 (0:00:00.040) 0:00:52.366 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 14:59:30 +0000 (0:00:00.028) 0:00:52.395 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/mapper/luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20", "fs_type": null }, { "action": "destroy format", "device": "/dev/nvme1n1", "fs_type": "luks" }, { "action": "create format", "device": "/dev/nvme1n1", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/nvme1n1", "name": "luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20", "password": "-", "state": "absent" } ], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "nvme1n1" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 14:59:30 +0000 (0:00:00.040) 0:00:52.435 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "nvme1n1" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 14:59:30 +0000 (0:00:00.037) 0:00:52.473 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 14:59:30 +0000 (0:00:00.040) 0:00:52.513 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 14:59:31 +0000 (0:00:00.338) 0:00:52.852 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 14:59:31 +0000 (0:00:00.457) 0:00:53.309 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 14:59:31 +0000 (0:00:00.361) 0:00:53.671 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 14:59:32 +0000 (0:00:00.447) 0:00:54.118 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658415554.8581505, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "a8111802e76fa3265aaef8b6e816500a5588bc5a", "ctime": 1658415552.7411506, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 8521508, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0600", "mtime": 1658415552.7411506, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 86, "uid": 0, "version": "18446744071955862706", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 14:59:32 +0000 (0:00:00.321) 0:00:54.440 ********* changed: [/cache/rhel-7.qcow2] => (item={'state': 'absent', 'password': '-', 'name': 'luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20', 'backing_device': '/dev/nvme1n1'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/nvme1n1", "name": "luks-a0dd669f-b9b3-4844-8d7f-f75db79f0d20", "password": "-", "state": "absent" }, "found": 1 } MSG: 1 line(s) removed TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 14:59:32 +0000 (0:00:00.361) 0:00:54.802 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:167 Thursday 21 July 2022 14:59:33 +0000 (0:00:00.874) 0:00:55.676 ********* included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1 Thursday 21 July 2022 14:59:33 +0000 (0:00:00.038) 0:00:55.715 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_pools_list": [ { "disks": [ "nvme1n1" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6 Thursday 21 July 2022 14:59:33 +0000 (0:00:00.048) 0:00:55.763 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14 Thursday 21 July 2022 14:59:33 +0000 (0:00:00.033) 0:00:55.797 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "4G", "type": "lvm", "uuid": "ec9aad2c-af48-4807-9d4d-f268ee3b75b7" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "LVM2_member", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "AxP00b-2jCE-LZk9-EuQd-Q9fq-J1B5-vXSV22" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-14-58-25-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19 Thursday 21 July 2022 14:59:34 +0000 (0:00:00.308) 0:00:56.105 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003406", "end": "2022-07-21 10:59:34.236051", "rc": 0, "start": "2022-07-21 10:59:34.232645" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24 Thursday 21 July 2022 14:59:34 +0000 (0:00:00.302) 0:00:56.408 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003363", "end": "2022-07-21 10:59:34.536360", "failed_when_result": false, "rc": 0, "start": "2022-07-21 10:59:34.532997" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33 Thursday 21 July 2022 14:59:34 +0000 (0:00:00.301) 0:00:56.710 ********* included: /tmp/tmpaxjje44y/tests/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': 'yabbadabbadoo', 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': 0, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': 'luks1', 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:5 Thursday 21 July 2022 14:59:34 +0000 (0:00:00.056) 0:00:56.767 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:18 Thursday 21 July 2022 14:59:34 +0000 (0:00:00.034) 0:00:56.801 ********* included: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members) included: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 14:59:35 +0000 (0:00:00.048) 0:00:56.849 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/nvme1n1" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 14:59:35 +0000 (0:00:00.056) 0:00:56.906 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/nvme1n1", "pv": "/dev/nvme1n1" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 14:59:35 +0000 (0:00:00.308) 0:00:57.215 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 14:59:35 +0000 (0:00:00.048) 0:00:57.263 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/nvme1n1" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 14:59:35 +0000 (0:00:00.086) 0:00:57.350 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 14:59:35 +0000 (0:00:00.055) 0:00:57.406 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 14:59:35 +0000 (0:00:00.040) 0:00:57.446 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 14:59:35 +0000 (0:00:00.086) 0:00:57.532 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 14:59:35 +0000 (0:00:00.027) 0:00:57.560 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/nvme1n1" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 14:59:35 +0000 (0:00:00.047) 0:00:57.607 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-md.yml for /cache/rhel-7.qcow2 TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:6 Thursday 21 July 2022 14:59:35 +0000 (0:00:00.042) 0:00:57.650 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:12 Thursday 21 July 2022 14:59:35 +0000 (0:00:00.024) 0:00:57.675 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:16 Thursday 21 July 2022 14:59:35 +0000 (0:00:00.025) 0:00:57.701 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:20 Thursday 21 July 2022 14:59:35 +0000 (0:00:00.026) 0:00:57.727 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:24 Thursday 21 July 2022 14:59:35 +0000 (0:00:00.028) 0:00:57.756 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:30 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.067) 0:00:57.824 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:36 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.026) 0:00:57.850 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:44 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.027) 0:00:57.878 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.037) 0:00:57.916 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.046) 0:00:57.962 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.043) 0:00:58.006 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.026) 0:00:58.033 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.025) 0:00:58.058 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.025) 0:00:58.084 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml for /cache/rhel-7.qcow2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.044) 0:00:58.129 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.044) 0:00:58.174 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.025) 0:00:58.199 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.024) 0:00:58.223 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.022) 0:00:58.246 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.022) 0:00:58.268 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2 TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.045) 0:00:58.314 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.047) 0:00:58.361 ********* skipping: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1) => { "_storage_test_pool_member_path": "/dev/nvme1n1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.025) 0:00:58.387 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/nvme1n1) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.042) 0:00:58.429 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.052) 0:00:58.482 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.049) 0:00:58.531 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.036) 0:00:58.568 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.037) 0:00:58.606 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.037) 0:00:58.644 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.039) 0:00:58.684 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.036) 0:00:58.720 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.048) 0:00:58.769 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 14:59:36 +0000 (0:00:00.047) 0:00:58.816 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.024) 0:00:58.840 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.023) 0:00:58.864 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.026) 0:00:58.890 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.025) 0:00:58.916 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.024) 0:00:58.941 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.023) 0:00:58.964 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.030) 0:00:58.995 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.104) 0:00:59.100 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.033) 0:00:59.134 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.041) 0:00:59.175 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.049) 0:00:59.225 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.080) 0:00:59.306 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.043) 0:00:59.350 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1037768, "block_size": 4096, "block_total": 1046016, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4250697728, "size_total": 4284481536, "uuid": "ec9aad2c-af48-4807-9d4d-f268ee3b75b7" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1037768, "block_size": 4096, "block_total": 1046016, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4250697728, "size_total": 4284481536, "uuid": "ec9aad2c-af48-4807-9d4d-f268ee3b75b7" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.059) 0:00:59.409 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.052) 0:00:59.462 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.050) 0:00:59.513 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.055) 0:00:59.568 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.025) 0:00:59.594 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.023) 0:00:59.618 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.023) 0:00:59.642 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.039) 0:00:59.681 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.062) 0:00:59.743 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 14:59:37 +0000 (0:00:00.052) 0:00:59.796 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 14:59:38 +0000 (0:00:00.053) 0:00:59.850 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 14:59:38 +0000 (0:00:00.036) 0:00:59.886 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 14:59:38 +0000 (0:00:00.033) 0:00:59.920 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 14:59:38 +0000 (0:00:00.038) 0:00:59.958 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 14:59:38 +0000 (0:00:00.042) 0:01:00.000 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658415570.1151507, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658415570.1151507, "dev": 5, "device_type": 64512, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 43481, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658415570.1151507, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 14:59:38 +0000 (0:00:00.323) 0:01:00.323 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 14:59:38 +0000 (0:00:00.038) 0:01:00.362 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 14:59:38 +0000 (0:00:00.039) 0:01:00.402 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 14:59:38 +0000 (0:00:00.039) 0:01:00.442 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 14:59:38 +0000 (0:00:00.026) 0:01:00.469 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 14:59:38 +0000 (0:00:00.040) 0:01:00.509 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 14:59:38 +0000 (0:00:00.025) 0:01:00.535 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.541) 0:01:01.076 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.024) 0:01:01.101 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.023) 0:01:01.124 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.051) 0:01:01.176 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.025) 0:01:01.201 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.025) 0:01:01.226 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.026) 0:01:01.253 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.025) 0:01:01.278 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.025) 0:01:01.303 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.053) 0:01:01.356 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.051) 0:01:01.408 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.035) 0:01:01.443 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.034) 0:01:01.478 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.034) 0:01:01.513 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.036) 0:01:01.549 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.037) 0:01:01.587 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.038) 0:01:01.625 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.036) 0:01:01.661 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.039) 0:01:01.701 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.048) 0:01:01.749 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 14:59:39 +0000 (0:00:00.042) 0:01:01.792 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 14:59:40 +0000 (0:00:00.038) 0:01:01.831 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 14:59:40 +0000 (0:00:00.310) 0:01:02.141 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 14:59:40 +0000 (0:00:00.332) 0:01:02.473 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "4294967296" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 14:59:40 +0000 (0:00:00.058) 0:01:02.532 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 14:59:40 +0000 (0:00:00.038) 0:01:02.570 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 14:59:40 +0000 (0:00:00.041) 0:01:02.611 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 14:59:40 +0000 (0:00:00.038) 0:01:02.650 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 14:59:40 +0000 (0:00:00.036) 0:01:02.687 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 14:59:40 +0000 (0:00:00.035) 0:01:02.723 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 14:59:40 +0000 (0:00:00.040) 0:01:02.763 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 14:59:40 +0000 (0:00:00.036) 0:01:02.800 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 14:59:41 +0000 (0:00:00.036) 0:01:02.836 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 14:59:41 +0000 (0:00:00.055) 0:01:02.892 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.022370", "end": "2022-07-21 10:59:41.048783", "rc": 0, "start": "2022-07-21 10:59:41.026413" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 14:59:41 +0000 (0:00:00.329) 0:01:03.221 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 14:59:41 +0000 (0:00:00.093) 0:01:03.315 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 14:59:41 +0000 (0:00:00.055) 0:01:03.371 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 14:59:41 +0000 (0:00:00.037) 0:01:03.408 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 14:59:41 +0000 (0:00:00.035) 0:01:03.444 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 14:59:41 +0000 (0:00:00.038) 0:01:03.482 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 14:59:41 +0000 (0:00:00.040) 0:01:03.522 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43 Thursday 21 July 2022 14:59:41 +0000 (0:00:00.090) 0:01:03.612 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53 Thursday 21 July 2022 14:59:41 +0000 (0:00:00.023) 0:01:03.636 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [create a file] *********************************************************** task path: /tmp/tmpaxjje44y/tests/create-test-file.yml:10 Thursday 21 July 2022 14:59:41 +0000 (0:00:00.035) 0:01:03.671 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "dest": "/opt/test1/quux", "gid": 0, "group": "root", "mode": "0644", "owner": "root", "secontext": "unconfined_u:object_r:unlabeled_t:s0", "size": 0, "state": "file", "uid": 0 } TASK [Add encryption to the pool] ********************************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:173 Thursday 21 July 2022 14:59:42 +0000 (0:00:00.327) 0:01:03.999 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 14:59:42 +0000 (0:00:00.039) 0:01:04.038 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 14:59:42 +0000 (0:00:00.034) 0:01:04.073 ********* ok: [/cache/rhel-7.qcow2] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 14:59:42 +0000 (0:00:00.412) 0:01:04.485 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 14:59:42 +0000 (0:00:00.064) 0:01:04.550 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 14:59:42 +0000 (0:00:00.035) 0:01:04.585 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 14:59:42 +0000 (0:00:00.035) 0:01:04.621 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 14:59:42 +0000 (0:00:00.046) 0:01:04.667 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 14:59:42 +0000 (0:00:00.019) 0:01:04.687 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 14:59:42 +0000 (0:00:00.028) 0:01:04.716 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": "serpent-xts-plain64", "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 14:59:42 +0000 (0:00:00.040) 0:01:04.756 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 14:59:42 +0000 (0:00:00.042) 0:01:04.798 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 14:59:43 +0000 (0:00:00.030) 0:01:04.829 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 14:59:43 +0000 (0:00:00.028) 0:01:04.857 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 14:59:43 +0000 (0:00:00.026) 0:01:04.883 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 14:59:43 +0000 (0:00:00.027) 0:01:04.911 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 14:59:43 +0000 (0:00:00.054) 0:01:04.965 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 14:59:43 +0000 (0:00:00.022) 0:01:04.988 ********* fatal: [/cache/rhel-7.qcow2]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: cannot remove and recreate existing pool 'foo' in safe mode TASK [linux-system-roles.storage : failed message] ***************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 21 July 2022 14:59:44 +0000 (0:00:01.052) 0:01:06.040 ********* fatal: [/cache/rhel-7.qcow2]: FAILED! => { "changed": false } MSG: {'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': True, 'pools': [{'name': 'foo', 'encryption_password': 'yabbadabbadoo', 'state': 'present', 'raid_metadata_version': None, 'encryption': True, 'encryption_key_size': 512, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': 'luks1', 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'fs_type': None, 'mount_options': None, 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': None, 'raid_level': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'deduplication': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'raid_disks': [], 'cache_mode': None, 'name': 'test1', 'cached': None, 'type': None, 'thin': False, 'cache_size': None, 'cache_devices': [], 'fs_create_options': None}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': 'serpent-xts-plain64', 'raid_spare_count': None}], 'volumes': [], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': [], 'msg': "cannot remove and recreate existing pool 'foo' in safe mode", '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 14:59:44 +0000 (0:00:00.079) 0:01:06.120 ********* TASK [Check that we failed in the role] **************************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:194 Thursday 21 July 2022 14:59:44 +0000 (0:00:00.023) 0:01:06.144 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the output of the safe_mode test] ********************************* task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:200 Thursday 21 July 2022 14:59:44 +0000 (0:00:00.124) 0:01:06.268 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [stat the file] *********************************************************** task path: /tmp/tmpaxjje44y/tests/verify-data-preservation.yml:10 Thursday 21 July 2022 14:59:44 +0000 (0:00:00.051) 0:01:06.320 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658415581.8221505, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658415581.8221505, "dev": 64512, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 67, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0644", "mtime": 1658415581.8221505, "nlink": 1, "path": "/opt/test1/quux", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "version": "18446744073475029210", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [assert file presence] **************************************************** task path: /tmp/tmpaxjje44y/tests/verify-data-preservation.yml:15 Thursday 21 July 2022 14:59:44 +0000 (0:00:00.327) 0:01:06.648 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Add encryption to the pool] ********************************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:211 Thursday 21 July 2022 14:59:44 +0000 (0:00:00.043) 0:01:06.691 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 14:59:44 +0000 (0:00:00.042) 0:01:06.734 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 14:59:44 +0000 (0:00:00.036) 0:01:06.770 ********* ok: [/cache/rhel-7.qcow2] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 14:59:45 +0000 (0:00:00.406) 0:01:07.176 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 14:59:45 +0000 (0:00:00.065) 0:01:07.242 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 14:59:45 +0000 (0:00:00.038) 0:01:07.280 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 14:59:45 +0000 (0:00:00.036) 0:01:07.316 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 14:59:45 +0000 (0:00:00.047) 0:01:07.364 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 14:59:45 +0000 (0:00:00.019) 0:01:07.384 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 14:59:45 +0000 (0:00:00.029) 0:01:07.413 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": "serpent-xts-plain64", "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 14:59:45 +0000 (0:00:00.040) 0:01:07.454 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 14:59:45 +0000 (0:00:00.039) 0:01:07.494 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 14:59:45 +0000 (0:00:00.031) 0:01:07.525 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 14:59:45 +0000 (0:00:00.026) 0:01:07.552 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 14:59:45 +0000 (0:00:00.027) 0:01:07.579 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 14:59:45 +0000 (0:00:00.031) 0:01:07.611 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 14:59:45 +0000 (0:00:00.058) 0:01:07.669 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 14:59:45 +0000 (0:00:00.023) 0:01:07.693 ********* changed: [/cache/rhel-7.qcow2] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/nvme1n1", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/nvme1n1", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/nvme1n1", "name": "luks-54a503f7-6643-449a-baac-c3779cfad410", "password": "-", "state": "present" } ], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "cryptsetup", "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": "serpent-xts-plain64", "encryption_key": null, "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 14:59:54 +0000 (0:00:08.163) 0:01:15.856 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 14:59:54 +0000 (0:00:00.037) 0:01:15.894 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 14:59:54 +0000 (0:00:00.023) 0:01:15.917 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/nvme1n1", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/nvme1n1", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/nvme1n1", "name": "luks-54a503f7-6643-449a-baac-c3779cfad410", "password": "-", "state": "present" } ], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "cryptsetup", "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": "serpent-xts-plain64", "encryption_key": null, "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 14:59:54 +0000 (0:00:00.041) 0:01:15.959 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": "serpent-xts-plain64", "encryption_key": null, "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 14:59:54 +0000 (0:00:00.039) 0:01:15.998 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 14:59:54 +0000 (0:00:00.038) 0:01:16.036 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 14:59:54 +0000 (0:00:00.348) 0:01:16.385 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 14:59:55 +0000 (0:00:00.487) 0:01:16.872 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 14:59:55 +0000 (0:00:00.360) 0:01:17.233 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 14:59:55 +0000 (0:00:00.446) 0:01:17.679 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658415574.5351505, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658415572.6211505, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 5088483, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658415572.6211505, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744072341144624", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 14:59:56 +0000 (0:00:00.318) 0:01:17.997 ********* changed: [/cache/rhel-7.qcow2] => (item={'state': 'present', 'password': '-', 'name': 'luks-54a503f7-6643-449a-baac-c3779cfad410', 'backing_device': '/dev/nvme1n1'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/nvme1n1", "name": "luks-54a503f7-6643-449a-baac-c3779cfad410", "password": "-", "state": "present" } } MSG: line added TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 14:59:56 +0000 (0:00:00.337) 0:01:18.335 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:230 Thursday 21 July 2022 14:59:58 +0000 (0:00:01.892) 0:01:20.227 ********* included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1 Thursday 21 July 2022 14:59:58 +0000 (0:00:00.044) 0:01:20.271 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_pools_list": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": "serpent-xts-plain64", "encryption_key": null, "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6 Thursday 21 July 2022 14:59:58 +0000 (0:00:00.058) 0:01:20.330 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14 Thursday 21 July 2022 14:59:58 +0000 (0:00:00.037) 0:01:20.367 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "4G", "type": "lvm", "uuid": "23dc4eb9-0913-4e77-81b3-832460b25c2f" }, "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410": { "fstype": "LVM2_member", "label": "", "name": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "size": "10G", "type": "crypt", "uuid": "NqWB4U-PJdY-nK6V-9YAV-ufOh-3OB3-AV93EB" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "crypto_LUKS", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "54a503f7-6643-449a-baac-c3779cfad410" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-14-58-25-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19 Thursday 21 July 2022 14:59:58 +0000 (0:00:00.362) 0:01:20.730 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003437", "end": "2022-07-21 10:59:58.865384", "rc": 0, "start": "2022-07-21 10:59:58.861947" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24 Thursday 21 July 2022 14:59:59 +0000 (0:00:00.310) 0:01:21.040 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003415", "end": "2022-07-21 10:59:59.181255", "failed_when_result": false, "rc": 0, "start": "2022-07-21 10:59:59.177840" } STDOUT: luks-54a503f7-6643-449a-baac-c3779cfad410 /dev/nvme1n1 - TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33 Thursday 21 July 2022 14:59:59 +0000 (0:00:00.347) 0:01:21.387 ********* included: /tmp/tmpaxjje44y/tests/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': 'yabbadabbadoo', 'state': 'present', 'raid_metadata_version': None, 'encryption': True, 'encryption_key_size': 512, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': 'luks1', 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': 'serpent-xts-plain64', 'raid_spare_count': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:5 Thursday 21 July 2022 14:59:59 +0000 (0:00:00.060) 0:01:21.448 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:18 Thursday 21 July 2022 14:59:59 +0000 (0:00:00.037) 0:01:21.485 ********* included: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members) included: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 14:59:59 +0000 (0:00:00.048) 0:01:21.533 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 14:59:59 +0000 (0:00:00.054) 0:01:21.588 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "pv": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.306) 0:01:21.895 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.048) 0:01:21.944 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.049) 0:01:21.993 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.051) 0:01:22.044 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "crypt" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.036) 0:01:22.081 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.023) 0:01:22.104 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.022) 0:01:22.127 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.044) 0:01:22.171 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-md.yml for /cache/rhel-7.qcow2 TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:6 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.041) 0:01:22.213 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:12 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.023) 0:01:22.236 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:16 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.023) 0:01:22.259 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:20 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.022) 0:01:22.282 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:24 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.025) 0:01:22.307 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:30 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.022) 0:01:22.330 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:36 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.023) 0:01:22.353 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:44 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.022) 0:01:22.375 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.033) 0:01:22.408 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.042) 0:01:22.451 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.040) 0:01:22.492 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.025) 0:01:22.517 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.027) 0:01:22.545 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.028) 0:01:22.574 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml for /cache/rhel-7.qcow2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.044) 0:01:22.618 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.041) 0:01:22.659 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.022) 0:01:22.682 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.022) 0:01:22.704 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.047) 0:01:22.752 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.023) 0:01:22.775 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2 TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 15:00:00 +0000 (0:00:00.045) 0:01:22.820 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "1", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 15:00:01 +0000 (0:00:00.048) 0:01:22.868 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml for /cache/rhel-7.qcow2 => (item=/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410) TASK [Get the backing device path] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:1 Thursday 21 July 2022 15:00:01 +0000 (0:00:00.042) 0:01:22.911 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "cmd": [ "realpath", "/dev/disk/by-uuid/54a503f7-6643-449a-baac-c3779cfad410" ], "delta": "0:00:00.003431", "end": "2022-07-21 11:00:01.061421", "rc": 0, "start": "2022-07-21 11:00:01.057990" } STDOUT: /dev/nvme1n1 TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:6 Thursday 21 July 2022 15:00:01 +0000 (0:00:00.325) 0:01:23.236 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this member] *************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:11 Thursday 21 July 2022 15:00:01 +0000 (0:00:00.543) 0:01:23.780 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cryptsetup", "luksDump", "/dev/nvme1n1" ], "delta": "0:00:00.039477", "end": "2022-07-21 11:00:01.974423", "rc": 0, "start": "2022-07-21 11:00:01.934946" } STDOUT: LUKS header information for /dev/nvme1n1 Version: 1 Cipher name: serpent Cipher mode: xts-plain64 Hash spec: sha256 Payload offset: 4096 MK bits: 512 MK digest: b9 e7 7d d7 56 75 da 2a 81 c2 44 fd 13 a1 f0 a7 0f 39 9d 8f MK salt: 7f 25 c8 0d 18 f4 42 28 04 e4 9d 54 f1 43 07 70 e8 09 bf 32 cc 44 6c f6 06 60 d2 72 cf 2c b4 94 MK iterations: 22946 UUID: 54a503f7-6643-449a-baac-c3779cfad410 Key Slot 0: ENABLED Iterations: 367148 Salt: 9d 15 56 23 73 7c 53 d6 ff 85 0d ff 75 6f 03 3b 3d ad 0f 24 7e ea 55 75 eb 94 3a d8 28 96 dc 13 Key material offset: 8 AF stripes: 4000 Key Slot 1: DISABLED Key Slot 2: DISABLED Key Slot 3: DISABLED Key Slot 4: DISABLED Key Slot 5: DISABLED Key Slot 6: DISABLED Key Slot 7: DISABLED TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:17 Thursday 21 July 2022 15:00:02 +0000 (0:00:00.373) 0:01:24.154 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:23 Thursday 21 July 2022 15:00:02 +0000 (0:00:00.055) 0:01:24.209 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:29 Thursday 21 July 2022 15:00:02 +0000 (0:00:00.056) 0:01:24.266 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 15:00:02 +0000 (0:00:00.052) 0:01:24.318 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 15:00:02 +0000 (0:00:00.046) 0:01:24.365 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [ "luks-54a503f7-6643-449a-baac-c3779cfad410 /dev/nvme1n1 -" ] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 15:00:02 +0000 (0:00:00.060) 0:01:24.425 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 15:00:02 +0000 (0:00:00.047) 0:01:24.473 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 15:00:02 +0000 (0:00:00.052) 0:01:24.525 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 15:00:02 +0000 (0:00:00.038) 0:01:24.564 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 15:00:02 +0000 (0:00:00.049) 0:01:24.613 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 15:00:02 +0000 (0:00:00.034) 0:01:24.648 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 15:00:02 +0000 (0:00:00.046) 0:01:24.694 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 15:00:02 +0000 (0:00:00.057) 0:01:24.752 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:00:02 +0000 (0:00:00.049) 0:01:24.801 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.071) 0:01:24.873 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.027) 0:01:24.901 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.026) 0:01:24.927 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.027) 0:01:24.954 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.025) 0:01:24.980 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.023) 0:01:25.004 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.025) 0:01:25.029 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.034) 0:01:25.064 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.035) 0:01:25.100 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.042) 0:01:25.142 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.051) 0:01:25.193 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.086) 0:01:25.280 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.044) 0:01:25.325 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1037768, "block_size": 4096, "block_total": 1046016, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4250697728, "size_total": 4284481536, "uuid": "23dc4eb9-0913-4e77-81b3-832460b25c2f" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1037768, "block_size": 4096, "block_total": 1046016, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4250697728, "size_total": 4284481536, "uuid": "23dc4eb9-0913-4e77-81b3-832460b25c2f" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.060) 0:01:25.385 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.056) 0:01:25.442 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.051) 0:01:25.493 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.055) 0:01:25.549 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.025) 0:01:25.575 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.026) 0:01:25.601 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.024) 0:01:25.626 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.036) 0:01:25.662 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.065) 0:01:25.727 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:00:03 +0000 (0:00:00.050) 0:01:25.777 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:00:04 +0000 (0:00:00.050) 0:01:25.828 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:00:04 +0000 (0:00:00.037) 0:01:25.865 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:00:04 +0000 (0:00:00.040) 0:01:25.906 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:00:04 +0000 (0:00:00.041) 0:01:25.947 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:00:04 +0000 (0:00:00.043) 0:01:25.990 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658415593.6441505, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658415593.6441505, "dev": 5, "device_type": 64513, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 46026, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658415593.6441505, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:00:04 +0000 (0:00:00.309) 0:01:26.300 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:00:04 +0000 (0:00:00.042) 0:01:26.342 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:00:04 +0000 (0:00:00.040) 0:01:26.383 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:00:04 +0000 (0:00:00.073) 0:01:26.457 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:00:04 +0000 (0:00:00.024) 0:01:26.482 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:00:04 +0000 (0:00:00.091) 0:01:26.573 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:00:04 +0000 (0:00:00.027) 0:01:26.601 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.526) 0:01:27.127 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.027) 0:01:27.154 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.025) 0:01:27.180 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.057) 0:01:27.237 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.025) 0:01:27.263 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.035) 0:01:27.298 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.029) 0:01:27.327 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.030) 0:01:27.358 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.030) 0:01:27.388 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.058) 0:01:27.447 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.055) 0:01:27.502 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.035) 0:01:27.538 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.038) 0:01:27.577 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.040) 0:01:27.617 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.041) 0:01:27.659 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.039) 0:01:27.699 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.037) 0:01:27.736 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.037) 0:01:27.774 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:00:05 +0000 (0:00:00.043) 0:01:27.817 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:00:06 +0000 (0:00:00.036) 0:01:27.853 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:00:06 +0000 (0:00:00.035) 0:01:27.888 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:00:06 +0000 (0:00:00.035) 0:01:27.924 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:00:06 +0000 (0:00:00.315) 0:01:28.239 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:00:06 +0000 (0:00:00.309) 0:01:28.548 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "4294967296" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:00:06 +0000 (0:00:00.053) 0:01:28.602 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:00:06 +0000 (0:00:00.040) 0:01:28.643 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:00:06 +0000 (0:00:00.039) 0:01:28.683 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:00:06 +0000 (0:00:00.049) 0:01:28.732 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:00:06 +0000 (0:00:00.036) 0:01:28.769 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:00:06 +0000 (0:00:00.035) 0:01:28.804 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:00:07 +0000 (0:00:00.038) 0:01:28.843 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:00:07 +0000 (0:00:00.034) 0:01:28.877 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:00:07 +0000 (0:00:00.094) 0:01:28.972 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:00:07 +0000 (0:00:00.054) 0:01:29.026 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.027079", "end": "2022-07-21 11:00:07.182068", "rc": 0, "start": "2022-07-21 11:00:07.154989" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:00:07 +0000 (0:00:00.331) 0:01:29.358 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:00:07 +0000 (0:00:00.110) 0:01:29.468 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:00:07 +0000 (0:00:00.135) 0:01:29.604 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:00:07 +0000 (0:00:00.042) 0:01:29.647 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:00:07 +0000 (0:00:00.037) 0:01:29.684 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:00:07 +0000 (0:00:00.039) 0:01:29.723 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:00:07 +0000 (0:00:00.040) 0:01:29.764 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43 Thursday 21 July 2022 15:00:07 +0000 (0:00:00.038) 0:01:29.802 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53 Thursday 21 July 2022 15:00:08 +0000 (0:00:00.024) 0:01:29.826 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [create a file] *********************************************************** task path: /tmp/tmpaxjje44y/tests/create-test-file.yml:10 Thursday 21 July 2022 15:00:08 +0000 (0:00:00.039) 0:01:29.865 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "dest": "/opt/test1/quux", "gid": 0, "group": "root", "mode": "0644", "owner": "root", "secontext": "unconfined_u:object_r:unlabeled_t:s0", "size": 0, "state": "file", "uid": 0 } TASK [Change the mountpoint, leaving encryption in place] ********************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:234 Thursday 21 July 2022 15:00:08 +0000 (0:00:00.319) 0:01:30.185 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 15:00:08 +0000 (0:00:00.049) 0:01:30.235 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 15:00:08 +0000 (0:00:00.038) 0:01:30.273 ********* ok: [/cache/rhel-7.qcow2] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 15:00:08 +0000 (0:00:00.417) 0:01:30.691 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 15:00:08 +0000 (0:00:00.063) 0:01:30.754 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 15:00:08 +0000 (0:00:00.033) 0:01:30.788 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 15:00:09 +0000 (0:00:00.033) 0:01:30.822 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 15:00:09 +0000 (0:00:00.046) 0:01:30.868 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 15:00:09 +0000 (0:00:00.019) 0:01:30.888 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 15:00:09 +0000 (0:00:00.026) 0:01:30.915 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "nvme1n1" ], "name": "foo", "type": "lvm", "volumes": [ { "mount_point": "/opt/test2", "name": "test1" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 15:00:09 +0000 (0:00:00.037) 0:01:30.952 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 15:00:09 +0000 (0:00:00.040) 0:01:30.992 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 15:00:09 +0000 (0:00:00.029) 0:01:31.022 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 15:00:09 +0000 (0:00:00.037) 0:01:31.059 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 15:00:09 +0000 (0:00:00.044) 0:01:31.104 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 15:00:09 +0000 (0:00:00.029) 0:01:31.133 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 15:00:09 +0000 (0:00:00.059) 0:01:31.193 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 15:00:09 +0000 (0:00:00.025) 0:01:31.218 ********* ok: [/cache/rhel-7.qcow2] => { "actions": [], "changed": false, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1" ], "mounts": [ { "path": "/opt/test1", "state": "absent" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "cryptsetup", "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "nvme1n1" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 4294967296, "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 15:00:10 +0000 (0:00:01.248) 0:01:32.467 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 15:00:10 +0000 (0:00:00.036) 0:01:32.503 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 15:00:10 +0000 (0:00:00.021) 0:01:32.524 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1" ], "mounts": [ { "path": "/opt/test1", "state": "absent" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "cryptsetup", "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "nvme1n1" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 4294967296, "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 15:00:10 +0000 (0:00:00.037) 0:01:32.562 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "nvme1n1" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 4294967296, "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 15:00:10 +0000 (0:00:00.038) 0:01:32.601 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 15:00:10 +0000 (0:00:00.033) 0:01:32.635 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'path': '/opt/test1', 'state': 'absent'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "mount_info": { "path": "/opt/test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 15:00:11 +0000 (0:00:00.351) 0:01:32.986 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 15:00:11 +0000 (0:00:00.468) 0:01:33.454 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test2', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 15:00:12 +0000 (0:00:01.358) 0:01:34.813 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 15:00:13 +0000 (0:00:00.447) 0:01:35.260 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658415599.1801505, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "142d9a6d6c28bb957805aa97c767fa6a32644bf9", "ctime": 1658415596.1531506, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 16777311, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0600", "mtime": 1658415596.1531506, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 57, "uid": 0, "version": "18446744073675325381", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 15:00:13 +0000 (0:00:00.312) 0:01:35.572 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 15:00:13 +0000 (0:00:00.022) 0:01:35.595 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [Assert to implicitly preserve encryption on existing pool] *************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:246 Thursday 21 July 2022 15:00:14 +0000 (0:00:00.829) 0:01:36.425 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [stat the file] *********************************************************** task path: /tmp/tmpaxjje44y/tests/verify-data-preservation.yml:10 Thursday 21 July 2022 15:00:14 +0000 (0:00:00.038) 0:01:36.463 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658415608.0061505, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658415608.0061505, "dev": 64513, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 67, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0644", "mtime": 1658415608.0061505, "nlink": 1, "path": "/opt/test2/quux", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "version": "18446744073073898251", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [assert file presence] **************************************************** task path: /tmp/tmpaxjje44y/tests/verify-data-preservation.yml:15 Thursday 21 July 2022 15:00:14 +0000 (0:00:00.316) 0:01:36.780 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:255 Thursday 21 July 2022 15:00:15 +0000 (0:00:00.076) 0:01:36.856 ********* included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1 Thursday 21 July 2022 15:00:15 +0000 (0:00:00.042) 0:01:36.899 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_pools_list": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "nvme1n1" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 4294967296, "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6 Thursday 21 July 2022 15:00:15 +0000 (0:00:00.082) 0:01:36.981 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14 Thursday 21 July 2022 15:00:15 +0000 (0:00:00.039) 0:01:37.021 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "4G", "type": "lvm", "uuid": "23dc4eb9-0913-4e77-81b3-832460b25c2f" }, "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410": { "fstype": "LVM2_member", "label": "", "name": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "size": "10G", "type": "crypt", "uuid": "NqWB4U-PJdY-nK6V-9YAV-ufOh-3OB3-AV93EB" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "crypto_LUKS", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "54a503f7-6643-449a-baac-c3779cfad410" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-14-58-25-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19 Thursday 21 July 2022 15:00:15 +0000 (0:00:00.330) 0:01:37.351 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003653", "end": "2022-07-21 11:00:15.486689", "rc": 0, "start": "2022-07-21 11:00:15.483036" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 /dev/mapper/foo-test1 /opt/test2 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24 Thursday 21 July 2022 15:00:15 +0000 (0:00:00.312) 0:01:37.664 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003595", "end": "2022-07-21 11:00:15.806360", "failed_when_result": false, "rc": 0, "start": "2022-07-21 11:00:15.802765" } STDOUT: luks-54a503f7-6643-449a-baac-c3779cfad410 /dev/nvme1n1 - TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33 Thursday 21 July 2022 15:00:16 +0000 (0:00:00.320) 0:01:37.985 ********* included: /tmp/tmpaxjje44y/tests/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': True, 'encryption_key_size': 0, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': 'luks1', 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': 4294967296, 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:5 Thursday 21 July 2022 15:00:16 +0000 (0:00:00.060) 0:01:38.045 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:18 Thursday 21 July 2022 15:00:16 +0000 (0:00:00.034) 0:01:38.079 ********* included: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members) included: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 15:00:16 +0000 (0:00:00.046) 0:01:38.125 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 15:00:16 +0000 (0:00:00.071) 0:01:38.197 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "pv": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 15:00:16 +0000 (0:00:00.328) 0:01:38.525 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 15:00:16 +0000 (0:00:00.053) 0:01:38.579 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 15:00:16 +0000 (0:00:00.051) 0:01:38.631 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 15:00:16 +0000 (0:00:00.052) 0:01:38.683 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "crypt" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 15:00:16 +0000 (0:00:00.039) 0:01:38.723 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 15:00:16 +0000 (0:00:00.024) 0:01:38.748 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 15:00:16 +0000 (0:00:00.022) 0:01:38.771 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 15:00:16 +0000 (0:00:00.044) 0:01:38.815 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-md.yml for /cache/rhel-7.qcow2 TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:6 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.044) 0:01:38.860 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:12 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.025) 0:01:38.885 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:16 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.027) 0:01:38.913 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:20 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.025) 0:01:38.938 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:24 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.035) 0:01:38.974 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:30 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.026) 0:01:39.001 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:36 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.034) 0:01:39.036 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:44 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.036) 0:01:39.072 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.039) 0:01:39.112 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.047) 0:01:39.160 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': 4294967296, 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.083) 0:01:39.244 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.028) 0:01:39.272 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.039) 0:01:39.312 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.031) 0:01:39.343 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml for /cache/rhel-7.qcow2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.044) 0:01:39.388 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': 4294967296, 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.045) 0:01:39.434 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.023) 0:01:39.457 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.023) 0:01:39.481 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.023) 0:01:39.505 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.025) 0:01:39.530 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2 TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.047) 0:01:39.578 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "1", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.051) 0:01:39.629 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml for /cache/rhel-7.qcow2 => (item=/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410) TASK [Get the backing device path] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:1 Thursday 21 July 2022 15:00:17 +0000 (0:00:00.045) 0:01:39.675 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "cmd": [ "realpath", "/dev/disk/by-uuid/54a503f7-6643-449a-baac-c3779cfad410" ], "delta": "0:00:00.003310", "end": "2022-07-21 11:00:17.822685", "rc": 0, "start": "2022-07-21 11:00:17.819375" } STDOUT: /dev/nvme1n1 TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:6 Thursday 21 July 2022 15:00:18 +0000 (0:00:00.320) 0:01:39.995 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this member] *************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:11 Thursday 21 July 2022 15:00:18 +0000 (0:00:00.531) 0:01:40.527 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cryptsetup", "luksDump", "/dev/nvme1n1" ], "delta": "0:00:00.037386", "end": "2022-07-21 11:00:18.701497", "rc": 0, "start": "2022-07-21 11:00:18.664111" } STDOUT: LUKS header information for /dev/nvme1n1 Version: 1 Cipher name: serpent Cipher mode: xts-plain64 Hash spec: sha256 Payload offset: 4096 MK bits: 512 MK digest: b9 e7 7d d7 56 75 da 2a 81 c2 44 fd 13 a1 f0 a7 0f 39 9d 8f MK salt: 7f 25 c8 0d 18 f4 42 28 04 e4 9d 54 f1 43 07 70 e8 09 bf 32 cc 44 6c f6 06 60 d2 72 cf 2c b4 94 MK iterations: 22946 UUID: 54a503f7-6643-449a-baac-c3779cfad410 Key Slot 0: ENABLED Iterations: 367148 Salt: 9d 15 56 23 73 7c 53 d6 ff 85 0d ff 75 6f 03 3b 3d ad 0f 24 7e ea 55 75 eb 94 3a d8 28 96 dc 13 Key material offset: 8 AF stripes: 4000 Key Slot 1: DISABLED Key Slot 2: DISABLED Key Slot 3: DISABLED Key Slot 4: DISABLED Key Slot 5: DISABLED Key Slot 6: DISABLED Key Slot 7: DISABLED TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:17 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.352) 0:01:40.879 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:23 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.050) 0:01:40.930 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-encryption.yml:29 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.023) 0:01:40.953 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.025) 0:01:40.979 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.048) 0:01:41.027 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [ "luks-54a503f7-6643-449a-baac-c3779cfad410 /dev/nvme1n1 -" ] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.087) 0:01:41.114 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.112) 0:01:41.227 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.050) 0:01:41.277 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.035) 0:01:41.313 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.049) 0:01:41.362 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.035) 0:01:41.397 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.032) 0:01:41.430 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.049) 0:01:41.479 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': 4294967296, 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.042) 0:01:41.522 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.022) 0:01:41.545 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.023) 0:01:41.569 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.022) 0:01:41.591 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.022) 0:01:41.614 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.022) 0:01:41.636 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.023) 0:01:41.660 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.022) 0:01:41.683 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.032) 0:01:41.716 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.033) 0:01:41.750 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': 4294967296, 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:00:19 +0000 (0:00:00.046) 0:01:41.796 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.058) 0:01:41.855 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.086) 0:01:41.942 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.044) 0:01:41.986 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1037768, "block_size": 4096, "block_total": 1046016, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097148, "inode_total": 2097152, "inode_used": 4, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4250697728, "size_total": 4284481536, "uuid": "23dc4eb9-0913-4e77-81b3-832460b25c2f" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1037768, "block_size": 4096, "block_total": 1046016, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097148, "inode_total": 2097152, "inode_used": 4, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4250697728, "size_total": 4284481536, "uuid": "23dc4eb9-0913-4e77-81b3-832460b25c2f" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.059) 0:01:42.045 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.053) 0:01:42.099 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.051) 0:01:42.150 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.049) 0:01:42.200 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.023) 0:01:42.224 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.022) 0:01:42.247 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.022) 0:01:42.270 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.031) 0:01:42.301 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.061) 0:01:42.363 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.052) 0:01:42.415 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.047) 0:01:42.462 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.033) 0:01:42.496 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.074) 0:01:42.570 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.101) 0:01:42.671 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:00:20 +0000 (0:00:00.043) 0:01:42.715 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658415593.6441505, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658415593.6441505, "dev": 5, "device_type": 64513, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 46026, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658415593.6441505, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:00:21 +0000 (0:00:00.333) 0:01:43.049 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:00:21 +0000 (0:00:00.042) 0:01:43.091 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:00:21 +0000 (0:00:00.041) 0:01:43.133 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:00:21 +0000 (0:00:00.038) 0:01:43.172 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:00:21 +0000 (0:00:00.026) 0:01:43.198 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:00:21 +0000 (0:00:00.045) 0:01:43.244 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:00:21 +0000 (0:00:00.024) 0:01:43.268 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:00:21 +0000 (0:00:00.517) 0:01:43.786 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:00:21 +0000 (0:00:00.024) 0:01:43.810 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.022) 0:01:43.833 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.053) 0:01:43.886 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.026) 0:01:43.913 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.025) 0:01:43.938 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.025) 0:01:43.963 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.023) 0:01:43.987 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.026) 0:01:44.013 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.054) 0:01:44.067 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.052) 0:01:44.120 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.039) 0:01:44.160 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.038) 0:01:44.198 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.044) 0:01:44.242 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.037) 0:01:44.280 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.040) 0:01:44.320 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.036) 0:01:44.357 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.039) 0:01:44.396 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.037) 0:01:44.434 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.041) 0:01:44.475 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.037) 0:01:44.512 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:00:22 +0000 (0:00:00.038) 0:01:44.551 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:00:23 +0000 (0:00:00.344) 0:01:44.896 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:00:23 +0000 (0:00:00.332) 0:01:45.228 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "4294967296" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:00:23 +0000 (0:00:00.115) 0:01:45.344 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:00:23 +0000 (0:00:00.038) 0:01:45.383 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:00:23 +0000 (0:00:00.036) 0:01:45.419 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:00:23 +0000 (0:00:00.036) 0:01:45.456 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:00:23 +0000 (0:00:00.035) 0:01:45.491 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:00:23 +0000 (0:00:00.037) 0:01:45.528 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:00:23 +0000 (0:00:00.040) 0:01:45.569 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:00:23 +0000 (0:00:00.044) 0:01:45.613 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:00:23 +0000 (0:00:00.040) 0:01:45.654 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:00:23 +0000 (0:00:00.054) 0:01:45.708 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.025424", "end": "2022-07-21 11:00:23.869032", "rc": 0, "start": "2022-07-21 11:00:23.843608" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:00:24 +0000 (0:00:00.336) 0:01:46.044 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:00:24 +0000 (0:00:00.052) 0:01:46.097 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:00:24 +0000 (0:00:00.053) 0:01:46.151 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:00:24 +0000 (0:00:00.037) 0:01:46.189 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:00:24 +0000 (0:00:00.037) 0:01:46.227 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:00:24 +0000 (0:00:00.038) 0:01:46.265 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:00:24 +0000 (0:00:00.038) 0:01:46.304 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43 Thursday 21 July 2022 15:00:24 +0000 (0:00:00.035) 0:01:46.339 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53 Thursday 21 July 2022 15:00:24 +0000 (0:00:00.021) 0:01:46.361 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Clean up] **************************************************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:257 Thursday 21 July 2022 15:00:24 +0000 (0:00:00.032) 0:01:46.394 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 15:00:24 +0000 (0:00:00.056) 0:01:46.451 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 15:00:24 +0000 (0:00:00.036) 0:01:46.487 ********* ok: [/cache/rhel-7.qcow2] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 15:00:25 +0000 (0:00:00.422) 0:01:46.910 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 15:00:25 +0000 (0:00:00.060) 0:01:46.970 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 15:00:25 +0000 (0:00:00.034) 0:01:47.004 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 15:00:25 +0000 (0:00:00.035) 0:01:47.039 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 15:00:25 +0000 (0:00:00.045) 0:01:47.085 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 15:00:25 +0000 (0:00:00.022) 0:01:47.108 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 15:00:25 +0000 (0:00:00.068) 0:01:47.176 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 15:00:25 +0000 (0:00:00.037) 0:01:47.213 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": [ { "disks": [ "nvme1n1" ], "name": "foo", "state": "absent", "type": "disk" } ] } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 15:00:25 +0000 (0:00:00.039) 0:01:47.253 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 15:00:25 +0000 (0:00:00.030) 0:01:47.284 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 15:00:25 +0000 (0:00:00.031) 0:01:47.315 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 15:00:25 +0000 (0:00:00.030) 0:01:47.345 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 15:00:25 +0000 (0:00:00.029) 0:01:47.375 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 15:00:25 +0000 (0:00:00.060) 0:01:47.436 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 15:00:25 +0000 (0:00:00.023) 0:01:47.460 ********* changed: [/cache/rhel-7.qcow2] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "fs_type": null }, { "action": "destroy format", "device": "/dev/nvme1n1", "fs_type": "luks" } ], "changed": true, "crypts": [ { "backing_device": "/dev/nvme1n1", "name": "luks-54a503f7-6643-449a-baac-c3779cfad410", "password": "-", "state": "absent" } ], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "xfsprogs" ], "pools": [], "volumes": [ { "_device": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "_mount_id": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "_raw_device": "/dev/nvme1n1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "lvmpv", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 10735321088, "state": "absent", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 15:00:27 +0000 (0:00:01.697) 0:01:49.158 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 15:00:27 +0000 (0:00:00.037) 0:01:49.195 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 15:00:27 +0000 (0:00:00.023) 0:01:49.219 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "fs_type": null }, { "action": "destroy format", "device": "/dev/nvme1n1", "fs_type": "luks" } ], "changed": true, "crypts": [ { "backing_device": "/dev/nvme1n1", "name": "luks-54a503f7-6643-449a-baac-c3779cfad410", "password": "-", "state": "absent" } ], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "xfsprogs" ], "pools": [], "volumes": [ { "_device": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "_mount_id": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "_raw_device": "/dev/nvme1n1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "lvmpv", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 10735321088, "state": "absent", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 15:00:27 +0000 (0:00:00.042) 0:01:49.261 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 15:00:27 +0000 (0:00:00.042) 0:01:49.304 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [ { "_device": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "_mount_id": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "_raw_device": "/dev/nvme1n1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "lvmpv", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 10735321088, "state": "absent", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 15:00:27 +0000 (0:00:00.047) 0:01:49.351 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'state': 'absent', 'path': '/opt/test2', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 15:00:27 +0000 (0:00:00.348) 0:01:49.699 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 15:00:28 +0000 (0:00:00.465) 0:01:50.164 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 15:00:28 +0000 (0:00:00.042) 0:01:50.207 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 15:00:28 +0000 (0:00:00.455) 0:01:50.663 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658415599.1801505, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "142d9a6d6c28bb957805aa97c767fa6a32644bf9", "ctime": 1658415596.1531506, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 16777311, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0600", "mtime": 1658415596.1531506, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 57, "uid": 0, "version": "18446744073675325381", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 15:00:29 +0000 (0:00:00.323) 0:01:50.986 ********* changed: [/cache/rhel-7.qcow2] => (item={'state': 'absent', 'password': '-', 'name': 'luks-54a503f7-6643-449a-baac-c3779cfad410', 'backing_device': '/dev/nvme1n1'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/nvme1n1", "name": "luks-54a503f7-6643-449a-baac-c3779cfad410", "password": "-", "state": "absent" }, "found": 1 } MSG: 1 line(s) removed TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 15:00:29 +0000 (0:00:00.331) 0:01:51.318 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/tests_luks_pool.yml:267 Thursday 21 July 2022 15:00:30 +0000 (0:00:00.875) 0:01:52.194 ********* included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1 Thursday 21 July 2022 15:00:30 +0000 (0:00:00.047) 0:01:52.241 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Print out volume information] ******************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6 Thursday 21 July 2022 15:00:30 +0000 (0:00:00.038) 0:01:52.280 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_volumes_list": [ { "_device": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "_mount_id": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410", "_raw_device": "/dev/nvme1n1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "lvmpv", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 10735321088, "state": "absent", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14 Thursday 21 July 2022 15:00:30 +0000 (0:00:00.102) 0:01:52.382 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-14-58-25-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19 Thursday 21 July 2022 15:00:30 +0000 (0:00:00.372) 0:01:52.754 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003369", "end": "2022-07-21 11:00:30.874240", "rc": 0, "start": "2022-07-21 11:00:30.870871" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24 Thursday 21 July 2022 15:00:31 +0000 (0:00:00.295) 0:01:53.050 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003727", "end": "2022-07-21 11:00:31.186120", "failed_when_result": false, "rc": 0, "start": "2022-07-21 11:00:31.182393" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33 Thursday 21 July 2022 15:00:31 +0000 (0:00:00.313) 0:01:53.364 ********* TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43 Thursday 21 July 2022 15:00:31 +0000 (0:00:00.024) 0:01:53.388 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/nvme1n1', 'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'lvmpv', 'mount_options': 'defaults', '_device': '/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410', 'size': 10735321088, 'mount_point': None, 'compression': None, 'encryption_password': None, 'encryption': True, 'raid_level': None, 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'type': 'disk', 'encryption_key_size': 0, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': 'luks1', 'cache_size': 0, '_mount_id': '/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410', 'raid_spare_count': None, 'name': 'foo', 'cache_mode': None, 'cache_devices': [], 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': None, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'thin_pool_size': None, 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:00:31 +0000 (0:00:00.059) 0:01:53.448 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:00:31 +0000 (0:00:00.052) 0:01:53.501 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:00:31 +0000 (0:00:00.078) 0:01:53.579 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/luks-54a503f7-6643-449a-baac-c3779cfad410" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:00:31 +0000 (0:00:00.053) 0:01:53.632 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:00:31 +0000 (0:00:00.056) 0:01:53.689 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:00:31 +0000 (0:00:00.023) 0:01:53.712 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:00:31 +0000 (0:00:00.047) 0:01:53.760 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:00:31 +0000 (0:00:00.038) 0:01:53.798 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.025) 0:01:53.823 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.024) 0:01:53.848 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.025) 0:01:53.874 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.040) 0:01:53.914 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.062) 0:01:53.977 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.023) 0:01:54.001 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.052) 0:01:54.053 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.036) 0:01:54.090 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.036) 0:01:54.127 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.025) 0:01:54.152 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.026) 0:01:54.179 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658415626.9491506, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658415626.9491506, "dev": 5, "device_type": 66306, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 9578, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/blockdevice", "mode": "0660", "mtime": 1658415626.9491506, "nlink": 1, "path": "/dev/nvme1n1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.321) 0:01:54.500 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.040) 0:01:54.541 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.024) 0:01:54.566 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "disk" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.077) 0:01:54.643 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.026) 0:01:54.669 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:00:32 +0000 (0:00:00.023) 0:01:54.692 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "exists": false } } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:00:33 +0000 (0:00:00.344) 0:01:55.036 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:00:33 +0000 (0:00:00.529) 0:01:55.566 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:00:33 +0000 (0:00:00.027) 0:01:55.593 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:00:33 +0000 (0:00:00.043) 0:01:55.637 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:00:33 +0000 (0:00:00.024) 0:01:55.661 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:00:33 +0000 (0:00:00.023) 0:01:55.685 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:00:33 +0000 (0:00:00.022) 0:01:55.707 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:00:33 +0000 (0:00:00.025) 0:01:55.733 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:00:33 +0000 (0:00:00.023) 0:01:55.756 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:00:33 +0000 (0:00:00.022) 0:01:55.779 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.049) 0:01:55.828 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.049) 0:01:55.877 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.033) 0:01:55.911 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.038) 0:01:55.949 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.037) 0:01:55.986 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.040) 0:01:56.027 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.038) 0:01:56.065 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.037) 0:01:56.102 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.037) 0:01:56.140 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.038) 0:01:56.178 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.036) 0:01:56.215 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.037) 0:01:56.252 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.038) 0:01:56.291 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.026) 0:01:56.318 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.037) 0:01:56.356 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.037) 0:01:56.394 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.036) 0:01:56.430 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.037) 0:01:56.468 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.038) 0:01:56.507 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.036) 0:01:56.543 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.035) 0:01:56.578 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.037) 0:01:56.615 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.036) 0:01:56.652 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.074) 0:01:56.727 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.025) 0:01:56.752 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.026) 0:01:56.779 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:00:34 +0000 (0:00:00.024) 0:01:56.803 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:00:35 +0000 (0:00:00.023) 0:01:56.827 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:00:35 +0000 (0:00:00.023) 0:01:56.851 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:00:35 +0000 (0:00:00.064) 0:01:56.916 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:00:35 +0000 (0:00:00.026) 0:01:56.942 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:00:35 +0000 (0:00:00.025) 0:01:56.968 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53 Thursday 21 July 2022 15:00:35 +0000 (0:00:00.037) 0:01:57.005 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* /cache/rhel-7.qcow2 : ok=543 changed=25 unreachable=0 failed=3 skipped=371 rescued=3 ignored=0 Thursday 21 July 2022 15:00:35 +0000 (0:00:00.050) 0:01:57.056 ********* =============================================================================== linux-system-roles.storage : make sure blivet is available -------------- 8.57s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 8.16s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 7.59s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 2.12s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : Update facts ------------------------------- 1.89s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.70s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : make sure required packages are installed --- 1.53s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 set up internal repositories -------------------------------------------- 1.42s /cache/rhel-7_setup.yml:5 ----------------------------------------------------- linux-system-roles.storage : set up new/current mounts ------------------ 1.36s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.25s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : get service facts -------------------------- 1.14s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.13s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.05s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Gathering Facts --------------------------------------------------------- 1.05s /tmp/tmpaxjje44y/tests/tests_luks_pool_nvme_generated.yml:3 ------------------- linux-system-roles.storage : get required packages ---------------------- 0.90s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 linux-system-roles.storage : Update facts ------------------------------- 0.88s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 linux-system-roles.storage : Update facts ------------------------------- 0.87s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 linux-system-roles.storage : Update facts ------------------------------- 0.87s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 0.87s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : Update facts ------------------------------- 0.83s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 ansible-playbook [core 2.12.6] config file = /etc/ansible/ansible.cfg configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.9/site-packages/ansible ansible collection location = /tmp/tmp5bkr4li_ executable location = /usr/bin/ansible-playbook python version = 3.9.13 (main, May 18 2022, 00:00:00) [GCC 11.3.1 20220421 (Red Hat 11.3.1-2)] jinja version = 2.11.3 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: rhel-7_setup.yml ***************************************************** 1 plays in /cache/rhel-7_setup.yml PLAY [Setup repos] ************************************************************* META: ran handlers TASK [set up internal repositories] ******************************************** task path: /cache/rhel-7_setup.yml:5 Thursday 21 July 2022 18:07:06 +0000 (0:00:00.018) 0:00:00.018 ********* changed: [/cache/rhel-7.qcow2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [/cache/rhel-7.qcow2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [/cache/rhel-7.qcow2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [/cache/rhel-7.qcow2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [/cache/rhel-7.qcow2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* /cache/rhel-7.qcow2 : ok=1 changed=1 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 Thursday 21 July 2022 18:07:07 +0000 (0:00:01.417) 0:00:01.435 ********* =============================================================================== set up internal repositories -------------------------------------------- 1.42s /cache/rhel-7_setup.yml:5 ----------------------------------------------------- statically imported: /tmp/tmptomayb7j/tests/storage/create-test-file.yml statically imported: /tmp/tmptomayb7j/tests/storage/verify-data-preservation.yml statically imported: /tmp/tmptomayb7j/tests/storage/create-test-file.yml statically imported: /tmp/tmptomayb7j/tests/storage/verify-data-preservation.yml statically imported: /tmp/tmptomayb7j/tests/storage/create-test-file.yml statically imported: /tmp/tmptomayb7j/tests/storage/verify-data-preservation.yml PLAYBOOK: tests_luks_pool_nvme_generated.yml *********************************** 2 plays in /tmp/tmptomayb7j/tests/storage/tests_luks_pool_nvme_generated.yml PLAY [all] ********************************************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool_nvme_generated.yml:3 Thursday 21 July 2022 18:07:07 +0000 (0:00:00.029) 0:00:01.464 ********* ok: [/cache/rhel-7.qcow2] META: ran handlers TASK [set disk interface for test] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool_nvme_generated.yml:7 Thursday 21 July 2022 18:07:08 +0000 (0:00:01.000) 0:00:02.465 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_use_interface": "nvme" }, "changed": false } META: ran handlers META: ran handlers PLAY [all] ********************************************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:2 Thursday 21 July 2022 18:07:08 +0000 (0:00:00.050) 0:00:02.515 ********* ok: [/cache/rhel-7.qcow2] META: ran handlers TASK [include_role : fedora.linux_system_roles.storage] ************************ task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:14 Thursday 21 July 2022 18:07:09 +0000 (0:00:00.739) 0:00:03.255 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 18:07:09 +0000 (0:00:00.036) 0:00:03.291 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 18:07:09 +0000 (0:00:00.028) 0:00:03.320 ********* ok: [/cache/rhel-7.qcow2] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 18:07:09 +0000 (0:00:00.424) 0:00:03.745 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 18:07:09 +0000 (0:00:00.053) 0:00:03.798 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 18:07:09 +0000 (0:00:00.028) 0:00:03.827 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 18:07:10 +0000 (0:00:00.028) 0:00:03.856 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 18:07:10 +0000 (0:00:00.052) 0:00:03.908 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 18:07:10 +0000 (0:00:00.016) 0:00:03.925 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "changes": { "installed": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "rc": 0, "results": [ "Loaded plugins: search-disabled-repos\nResolving Dependencies\n--> Running transaction check\n---> Package libblockdev-crypto.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libblockdev-utils(x86-64) = 2.18-5.el7 for package: libblockdev-crypto-2.18-5.el7.x86_64\n--> Processing Dependency: libvolume_key.so.1()(64bit) for package: libblockdev-crypto-2.18-5.el7.x86_64\n--> Processing Dependency: libbd_utils.so.2()(64bit) for package: libblockdev-crypto-2.18-5.el7.x86_64\n---> Package libblockdev-dm.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libdmraid.so.1(Base)(64bit) for package: libblockdev-dm-2.18-5.el7.x86_64\n--> Processing Dependency: dmraid for package: libblockdev-dm-2.18-5.el7.x86_64\n--> Processing Dependency: libdmraid.so.1()(64bit) for package: libblockdev-dm-2.18-5.el7.x86_64\n---> Package libblockdev-lvm.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: lvm2 for package: libblockdev-lvm-2.18-5.el7.x86_64\n--> Processing Dependency: device-mapper-persistent-data for package: libblockdev-lvm-2.18-5.el7.x86_64\n---> Package libblockdev-mdraid.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: mdadm for package: libblockdev-mdraid-2.18-5.el7.x86_64\n--> Processing Dependency: libbytesize.so.1()(64bit) for package: libblockdev-mdraid-2.18-5.el7.x86_64\n---> Package libblockdev-swap.x86_64 0:2.18-5.el7 will be installed\n---> Package python-enum34.noarch 0:1.0.4-1.el7 will be installed\n---> Package python2-blivet3.noarch 1:3.1.3-3.el7 will be installed\n--> Processing Dependency: blivet3-data = 1:3.1.3-3.el7 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-bytesize >= 0.3 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-blockdev >= 2.17 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: pyparted >= 3.9 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-hawkey for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: lsof for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Running transaction check\n---> Package blivet3-data.noarch 1:3.1.3-3.el7 will be installed\n---> Package device-mapper-persistent-data.x86_64 0:0.8.5-3.el7_9.2 will be installed\n--> Processing Dependency: libaio.so.1(LIBAIO_0.4)(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n--> Processing Dependency: libaio.so.1(LIBAIO_0.1)(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n--> Processing Dependency: libaio.so.1()(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n---> Package dmraid.x86_64 0:1.0.0.rc16-28.el7 will be installed\n--> Processing Dependency: libdevmapper-event.so.1.02(Base)(64bit) for package: dmraid-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: dmraid-events for package: dmraid-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: libdevmapper-event.so.1.02()(64bit) for package: dmraid-1.0.0.rc16-28.el7.x86_64\n---> Package libblockdev-utils.x86_64 0:2.18-5.el7 will be installed\n---> Package libbytesize.x86_64 0:1.2-1.el7 will be installed\n--> Processing Dependency: libmpfr.so.4()(64bit) for package: libbytesize-1.2-1.el7.x86_64\n---> Package lsof.x86_64 0:4.87-6.el7 will be installed\n---> Package lvm2.x86_64 7:2.02.187-6.el7_9.5 will be installed\n--> Processing Dependency: lvm2-libs = 7:2.02.187-6.el7_9.5 for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n--> Processing Dependency: liblvm2app.so.2.2(Base)(64bit) for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n--> Processing Dependency: liblvm2app.so.2.2()(64bit) for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n---> Package mdadm.x86_64 0:4.1-9.el7_9 will be installed\n--> Processing Dependency: libreport-filesystem for package: mdadm-4.1-9.el7_9.x86_64\n---> Package pyparted.x86_64 1:3.9-15.el7 will be installed\n---> Package python2-blockdev.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libblockdev(x86-64) = 2.18-5.el7 for package: python2-blockdev-2.18-5.el7.x86_64\n---> Package python2-bytesize.x86_64 0:1.2-1.el7 will be installed\n---> Package python2-hawkey.x86_64 0:0.22.5-2.el7_9 will be installed\n--> Processing Dependency: libdnf(x86-64) = 0.22.5-2.el7_9 for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: python2-libdnf = 0.22.5-2.el7_9 for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolv.so.0(SOLV_1.0)(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolvext.so.0(SOLV_1.0)(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libdnf.so.2()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libjson-glib-1.0.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libmodulemd.so.1()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: librepo.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: librhsm.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolv.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolvext.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n---> Package volume_key-libs.x86_64 0:0.3.9-9.el7 will be installed\n--> Running transaction check\n---> Package device-mapper-event-libs.x86_64 7:1.02.170-6.el7_9.5 will be installed\n---> Package dmraid-events.x86_64 0:1.0.0.rc16-28.el7 will be installed\n--> Processing Dependency: sgpio for package: dmraid-events-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: device-mapper-event for package: dmraid-events-1.0.0.rc16-28.el7.x86_64\n---> Package json-glib.x86_64 0:1.4.2-2.el7 will be installed\n---> Package libaio.x86_64 0:0.3.109-13.el7 will be installed\n---> Package libblockdev.x86_64 0:2.18-5.el7 will be installed\n---> Package libdnf.x86_64 0:0.22.5-2.el7_9 will be installed\n---> Package libmodulemd.x86_64 0:1.6.3-1.el7 will be installed\n---> Package librepo.x86_64 0:1.8.1-8.el7_9 will be installed\n---> Package libreport-filesystem.x86_64 0:2.1.11-53.el7 will be installed\n---> Package librhsm.x86_64 0:0.0.3-3.el7_9 will be installed\n---> Package libsolv.x86_64 0:0.6.34-4.el7 will be installed\n---> Package lvm2-libs.x86_64 7:2.02.187-6.el7_9.5 will be installed\n---> Package mpfr.x86_64 0:3.1.1-4.el7 will be installed\n---> Package python2-libdnf.x86_64 0:0.22.5-2.el7_9 will be installed\n--> Running transaction check\n---> Package device-mapper-event.x86_64 7:1.02.170-6.el7_9.5 will be installed\n---> Package sgpio.x86_64 0:1.2.0.10-13.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package Arch Version Repository Size\n================================================================================\nInstalling:\n libblockdev-crypto x86_64 2.18-5.el7 rhel 60 k\n libblockdev-dm x86_64 2.18-5.el7 rhel-optional 54 k\n libblockdev-lvm x86_64 2.18-5.el7 rhel 62 k\n libblockdev-mdraid x86_64 2.18-5.el7 rhel 57 k\n libblockdev-swap x86_64 2.18-5.el7 rhel 52 k\n python-enum34 noarch 1.0.4-1.el7 rhel 52 k\n python2-blivet3 noarch 1:3.1.3-3.el7 rhel 851 k\nInstalling for dependencies:\n blivet3-data noarch 1:3.1.3-3.el7 rhel 77 k\n device-mapper-event\n x86_64 7:1.02.170-6.el7_9.5 rhel 192 k\n device-mapper-event-libs\n x86_64 7:1.02.170-6.el7_9.5 rhel 192 k\n device-mapper-persistent-data\n x86_64 0.8.5-3.el7_9.2 rhel 423 k\n dmraid x86_64 1.0.0.rc16-28.el7 rhel 151 k\n dmraid-events x86_64 1.0.0.rc16-28.el7 rhel 21 k\n json-glib x86_64 1.4.2-2.el7 rhel 134 k\n libaio x86_64 0.3.109-13.el7 rhel 24 k\n libblockdev x86_64 2.18-5.el7 rhel 119 k\n libblockdev-utils x86_64 2.18-5.el7 rhel 59 k\n libbytesize x86_64 1.2-1.el7 rhel 52 k\n libdnf x86_64 0.22.5-2.el7_9 rhel-7-server-extras-rpms 536 k\n libmodulemd x86_64 1.6.3-1.el7 rhel-7-server-extras-rpms 153 k\n librepo x86_64 1.8.1-8.el7_9 rhel 82 k\n libreport-filesystem\n x86_64 2.1.11-53.el7 rhel 41 k\n librhsm x86_64 0.0.3-3.el7_9 rhel-7-server-extras-rpms 28 k\n libsolv x86_64 0.6.34-4.el7 rhel 329 k\n lsof x86_64 4.87-6.el7 rhel 331 k\n lvm2 x86_64 7:2.02.187-6.el7_9.5 rhel 1.3 M\n lvm2-libs x86_64 7:2.02.187-6.el7_9.5 rhel 1.1 M\n mdadm x86_64 4.1-9.el7_9 rhel 440 k\n mpfr x86_64 3.1.1-4.el7 rhel 203 k\n pyparted x86_64 1:3.9-15.el7 rhel 195 k\n python2-blockdev x86_64 2.18-5.el7 rhel 61 k\n python2-bytesize x86_64 1.2-1.el7 rhel 22 k\n python2-hawkey x86_64 0.22.5-2.el7_9 rhel-7-server-extras-rpms 71 k\n python2-libdnf x86_64 0.22.5-2.el7_9 rhel-7-server-extras-rpms 611 k\n sgpio x86_64 1.2.0.10-13.el7 rhel 14 k\n volume_key-libs x86_64 0.3.9-9.el7 rhel 141 k\n\nTransaction Summary\n================================================================================\nInstall 7 Packages (+29 Dependent packages)\n\nTotal download size: 8.2 M\nInstalled size: 24 M\nDownloading packages:\n--------------------------------------------------------------------------------\nTotal 18 MB/s | 8.2 MB 00:00 \nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n Installing : libblockdev-utils-2.18-5.el7.x86_64 1/36 \n Installing : 7:device-mapper-event-libs-1.02.170-6.el7_9.5.x86_64 2/36 \n Installing : json-glib-1.4.2-2.el7.x86_64 3/36 \n Installing : librhsm-0.0.3-3.el7_9.x86_64 4/36 \n Installing : libsolv-0.6.34-4.el7.x86_64 5/36 \n Installing : libaio-0.3.109-13.el7.x86_64 6/36 \n Installing : librepo-1.8.1-8.el7_9.x86_64 7/36 \n Installing : libmodulemd-1.6.3-1.el7.x86_64 8/36 \n Installing : libdnf-0.22.5-2.el7_9.x86_64 9/36 \n Installing : device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64 10/36 \n Installing : 7:device-mapper-event-1.02.170-6.el7_9.5.x86_64 11/36 \n Installing : 7:lvm2-libs-2.02.187-6.el7_9.5.x86_64 12/36 \n Installing : 7:lvm2-2.02.187-6.el7_9.5.x86_64 13/36 \n Installing : python2-libdnf-0.22.5-2.el7_9.x86_64 14/36 \n Installing : python2-hawkey-0.22.5-2.el7_9.x86_64 15/36 \n Installing : libblockdev-2.18-5.el7.x86_64 16/36 \n Installing : python2-blockdev-2.18-5.el7.x86_64 17/36 \n Installing : 1:pyparted-3.9-15.el7.x86_64 18/36 \n Installing : sgpio-1.2.0.10-13.el7.x86_64 19/36 \n Installing : dmraid-1.0.0.rc16-28.el7.x86_64 20/36 \n Installing : dmraid-events-1.0.0.rc16-28.el7.x86_64 21/36 \n Installing : volume_key-libs-0.3.9-9.el7.x86_64 22/36 \n Installing : mpfr-3.1.1-4.el7.x86_64 23/36 \n Installing : libbytesize-1.2-1.el7.x86_64 24/36 \n Installing : python2-bytesize-1.2-1.el7.x86_64 25/36 \n Installing : libreport-filesystem-2.1.11-53.el7.x86_64 26/36 \n Installing : mdadm-4.1-9.el7_9.x86_64 27/36 \n Installing : 1:blivet3-data-3.1.3-3.el7.noarch 28/36 \n Installing : lsof-4.87-6.el7.x86_64 29/36 \n Installing : 1:python2-blivet3-3.1.3-3.el7.noarch 30/36 \n Installing : libblockdev-mdraid-2.18-5.el7.x86_64 31/36 \n Installing : libblockdev-crypto-2.18-5.el7.x86_64 32/36 \n Installing : libblockdev-dm-2.18-5.el7.x86_64 33/36 \n Installing : libblockdev-lvm-2.18-5.el7.x86_64 34/36 \n Installing : libblockdev-swap-2.18-5.el7.x86_64 35/36 \n Installing : python-enum34-1.0.4-1.el7.noarch 36/36 \n Verifying : 7:device-mapper-event-1.02.170-6.el7_9.5.x86_64 1/36 \n Verifying : libblockdev-swap-2.18-5.el7.x86_64 2/36 \n Verifying : librhsm-0.0.3-3.el7_9.x86_64 3/36 \n Verifying : libblockdev-lvm-2.18-5.el7.x86_64 4/36 \n Verifying : lsof-4.87-6.el7.x86_64 5/36 \n Verifying : libblockdev-mdraid-2.18-5.el7.x86_64 6/36 \n Verifying : libdnf-0.22.5-2.el7_9.x86_64 7/36 \n Verifying : python-enum34-1.0.4-1.el7.noarch 8/36 \n Verifying : 1:blivet3-data-3.1.3-3.el7.noarch 9/36 \n Verifying : dmraid-events-1.0.0.rc16-28.el7.x86_64 10/36 \n Verifying : python2-blockdev-2.18-5.el7.x86_64 11/36 \n Verifying : libmodulemd-1.6.3-1.el7.x86_64 12/36 \n Verifying : librepo-1.8.1-8.el7_9.x86_64 13/36 \n Verifying : libblockdev-dm-2.18-5.el7.x86_64 14/36 \n Verifying : json-glib-1.4.2-2.el7.x86_64 15/36 \n Verifying : libaio-0.3.109-13.el7.x86_64 16/36 \n Verifying : 7:lvm2-libs-2.02.187-6.el7_9.5.x86_64 17/36 \n Verifying : python2-hawkey-0.22.5-2.el7_9.x86_64 18/36 \n Verifying : python2-bytesize-1.2-1.el7.x86_64 19/36 \n Verifying : libblockdev-2.18-5.el7.x86_64 20/36 \n Verifying : libreport-filesystem-2.1.11-53.el7.x86_64 21/36 \n Verifying : libbytesize-1.2-1.el7.x86_64 22/36 \n Verifying : 7:device-mapper-event-libs-1.02.170-6.el7_9.5.x86_64 23/36 \n Verifying : python2-libdnf-0.22.5-2.el7_9.x86_64 24/36 \n Verifying : 7:lvm2-2.02.187-6.el7_9.5.x86_64 25/36 \n Verifying : libblockdev-utils-2.18-5.el7.x86_64 26/36 \n Verifying : mpfr-3.1.1-4.el7.x86_64 27/36 \n Verifying : volume_key-libs-0.3.9-9.el7.x86_64 28/36 \n Verifying : libsolv-0.6.34-4.el7.x86_64 29/36 \n Verifying : device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64 30/36 \n Verifying : 1:python2-blivet3-3.1.3-3.el7.noarch 31/36 \n Verifying : dmraid-1.0.0.rc16-28.el7.x86_64 32/36 \n Verifying : mdadm-4.1-9.el7_9.x86_64 33/36 \n Verifying : sgpio-1.2.0.10-13.el7.x86_64 34/36 \n Verifying : libblockdev-crypto-2.18-5.el7.x86_64 35/36 \n Verifying : 1:pyparted-3.9-15.el7.x86_64 36/36 \n\nInstalled:\n libblockdev-crypto.x86_64 0:2.18-5.el7 libblockdev-dm.x86_64 0:2.18-5.el7 \n libblockdev-lvm.x86_64 0:2.18-5.el7 libblockdev-mdraid.x86_64 0:2.18-5.el7\n libblockdev-swap.x86_64 0:2.18-5.el7 python-enum34.noarch 0:1.0.4-1.el7 \n python2-blivet3.noarch 1:3.1.3-3.el7 \n\nDependency Installed:\n blivet3-data.noarch 1:3.1.3-3.el7 \n device-mapper-event.x86_64 7:1.02.170-6.el7_9.5 \n device-mapper-event-libs.x86_64 7:1.02.170-6.el7_9.5 \n device-mapper-persistent-data.x86_64 0:0.8.5-3.el7_9.2 \n dmraid.x86_64 0:1.0.0.rc16-28.el7 \n dmraid-events.x86_64 0:1.0.0.rc16-28.el7 \n json-glib.x86_64 0:1.4.2-2.el7 \n libaio.x86_64 0:0.3.109-13.el7 \n libblockdev.x86_64 0:2.18-5.el7 \n libblockdev-utils.x86_64 0:2.18-5.el7 \n libbytesize.x86_64 0:1.2-1.el7 \n libdnf.x86_64 0:0.22.5-2.el7_9 \n libmodulemd.x86_64 0:1.6.3-1.el7 \n librepo.x86_64 0:1.8.1-8.el7_9 \n libreport-filesystem.x86_64 0:2.1.11-53.el7 \n librhsm.x86_64 0:0.0.3-3.el7_9 \n libsolv.x86_64 0:0.6.34-4.el7 \n lsof.x86_64 0:4.87-6.el7 \n lvm2.x86_64 7:2.02.187-6.el7_9.5 \n lvm2-libs.x86_64 7:2.02.187-6.el7_9.5 \n mdadm.x86_64 0:4.1-9.el7_9 \n mpfr.x86_64 0:3.1.1-4.el7 \n pyparted.x86_64 1:3.9-15.el7 \n python2-blockdev.x86_64 0:2.18-5.el7 \n python2-bytesize.x86_64 0:1.2-1.el7 \n python2-hawkey.x86_64 0:0.22.5-2.el7_9 \n python2-libdnf.x86_64 0:0.22.5-2.el7_9 \n sgpio.x86_64 0:1.2.0.10-13.el7 \n volume_key-libs.x86_64 0:0.3.9-9.el7 \n\nComplete!\n" ] } TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 18:07:19 +0000 (0:00:08.995) 0:00:12.920 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 18:07:19 +0000 (0:00:00.080) 0:00:13.000 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 18:07:19 +0000 (0:00:00.059) 0:00:13.059 ********* ok: [/cache/rhel-7.qcow2] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 18:07:19 +0000 (0:00:00.618) 0:00:13.678 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 18:07:19 +0000 (0:00:00.066) 0:00:13.745 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 18:07:19 +0000 (0:00:00.030) 0:00:13.775 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 18:07:19 +0000 (0:00:00.034) 0:00:13.810 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 18:07:19 +0000 (0:00:00.031) 0:00:13.842 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [] } TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 18:07:20 +0000 (0:00:00.521) 0:00:14.364 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "brandbot.service": { "name": "brandbot.service", "source": "systemd", "state": "stopped", "status": "static" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-shell.service": { "name": "console-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.import1.service": { "name": "dbus-org.freedesktop.import1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.machine1.service": { "name": "dbus-org.freedesktop.machine1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dmraid-activation.service": { "name": "dmraid-activation.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "exim.service": { "name": "exim.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "ip6tables.service": { "name": "ip6tables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iptables.service": { "name": "iptables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmetad.service": { "name": "lvm2-lvmetad.service", "source": "systemd", "state": "running", "status": "static" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "netconsole": { "name": "netconsole", "source": "sysv", "state": "stopped", "status": "disabled" }, "network": { "name": "network", "source": "sysv", "state": "running", "status": "enabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "active" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-config.service": { "name": "nfs-config.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-idmap.service": { "name": "nfs-idmap.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-lock.service": { "name": "nfs-lock.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-rquotad.service": { "name": "nfs-rquotad.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-secure-server.service": { "name": "nfs-secure-server.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-secure.service": { "name": "nfs-secure.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs.service": { "name": "nfs.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfslock.service": { "name": "nfslock.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ovirt-guest-agent.service": { "name": "ovirt-guest-agent.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "postfix.service": { "name": "postfix.service", "source": "systemd", "state": "running", "status": "enabled" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhel-autorelabel-mark.service": { "name": "rhel-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-autorelabel.service": { "name": "rhel-autorelabel.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-configure.service": { "name": "rhel-configure.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-dmesg.service": { "name": "rhel-dmesg.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-domainname.service": { "name": "rhel-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-import-state.service": { "name": "rhel-import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-loadmodules.service": { "name": "rhel-loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-readonly.service": { "name": "rhel-readonly.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhnsd": { "name": "rhnsd", "source": "sysv", "state": "running", "status": "enabled" }, "rhnsd.service": { "name": "rhnsd.service", "source": "systemd", "state": "running", "status": "active" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-rquotad.service": { "name": "rpc-rquotad.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpcgssd.service": { "name": "rpcgssd.service", "source": "systemd", "state": "inactive", "status": "static" }, "rpcidmapd.service": { "name": "rpcidmapd.service", "source": "systemd", "state": "inactive", "status": "static" }, "rsyncd.service": { "name": "rsyncd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyncd@.service": { "name": "rsyncd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-policy-migrate-local-changes@.service": { "name": "selinux-policy-migrate-local-changes@.service", "source": "systemd", "state": "unknown", "status": "static" }, "selinux-policy-migrate-local-changes@targeted.service": { "name": "selinux-policy-migrate-local-changes@targeted.service", "source": "systemd", "state": "stopped", "status": "active" }, "sendmail.service": { "name": "sendmail.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "static" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bootchart.service": { "name": "systemd-bootchart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-importd.service": { "name": "systemd-importd.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machined.service": { "name": "systemd-machined.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-nspawn@.service": { "name": "systemd-nspawn@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-readahead-collect.service": { "name": "systemd-readahead-collect.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-readahead-done.service": { "name": "systemd-readahead-done.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "systemd-readahead-drop.service": { "name": "systemd-readahead-drop.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "systemd-readahead-replay.service": { "name": "systemd-readahead-replay.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill@.service": { "name": "systemd-rfkill@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-shutdownd.service": { "name": "systemd-shutdownd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "wpa_supplicant.service": { "name": "wpa_supplicant.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 18:07:21 +0000 (0:00:01.105) 0:00:15.469 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 18:07:21 +0000 (0:00:00.054) 0:00:15.524 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 18:07:21 +0000 (0:00:00.020) 0:00:15.545 ********* ok: [/cache/rhel-7.qcow2] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 18:07:22 +0000 (0:00:00.460) 0:00:16.005 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 18:07:22 +0000 (0:00:00.034) 0:00:16.040 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 18:07:22 +0000 (0:00:00.020) 0:00:16.060 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 18:07:22 +0000 (0:00:00.034) 0:00:16.095 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 18:07:22 +0000 (0:00:00.033) 0:00:16.128 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 18:07:22 +0000 (0:00:00.032) 0:00:16.161 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 18:07:22 +0000 (0:00:00.037) 0:00:16.198 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 18:07:22 +0000 (0:00:00.022) 0:00:16.221 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 18:07:22 +0000 (0:00:00.033) 0:00:16.254 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 18:07:22 +0000 (0:00:00.021) 0:00:16.276 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658426836.3731558, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658201031.524, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 70, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658200515.884, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744071677828413", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 18:07:22 +0000 (0:00:00.416) 0:00:16.692 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 18:07:22 +0000 (0:00:00.020) 0:00:16.713 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [Mark tasks to be skipped] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:17 Thursday 21 July 2022 18:07:23 +0000 (0:00:00.836) 0:00:17.549 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_skip_checks": [ "blivet_available", "service_facts" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:23 Thursday 21 July 2022 18:07:23 +0000 (0:00:00.034) 0:00:17.583 ********* included: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml for /cache/rhel-7.qcow2 TASK [Find unused disks in the system] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml:2 Thursday 21 July 2022 18:07:23 +0000 (0:00:00.035) 0:00:17.619 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "disks": [ "nvme1n1" ] } TASK [Set unused_disks if necessary] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml:9 Thursday 21 July 2022 18:07:24 +0000 (0:00:00.482) 0:00:18.101 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "unused_disks": [ "nvme1n1" ] }, "changed": false } TASK [Exit playbook when there's not enough unused disks in the system] ******** task path: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml:14 Thursday 21 July 2022 18:07:24 +0000 (0:00:00.035) 0:00:18.136 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Print unused disks] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml:19 Thursday 21 July 2022 18:07:24 +0000 (0:00:00.039) 0:00:18.175 ********* ok: [/cache/rhel-7.qcow2] => { "unused_disks": [ "nvme1n1" ] } TASK [Create an encrypted lvm pool] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:34 Thursday 21 July 2022 18:07:24 +0000 (0:00:00.034) 0:00:18.209 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 18:07:24 +0000 (0:00:00.043) 0:00:18.253 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 18:07:24 +0000 (0:00:00.034) 0:00:18.288 ********* ok: [/cache/rhel-7.qcow2] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 18:07:24 +0000 (0:00:00.410) 0:00:18.699 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 18:07:24 +0000 (0:00:00.061) 0:00:18.760 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 18:07:24 +0000 (0:00:00.033) 0:00:18.794 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 18:07:24 +0000 (0:00:00.036) 0:00:18.830 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 18:07:25 +0000 (0:00:00.054) 0:00:18.885 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 18:07:25 +0000 (0:00:00.019) 0:00:18.904 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 18:07:25 +0000 (0:00:00.026) 0:00:18.931 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "name": "foo", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 18:07:25 +0000 (0:00:00.041) 0:00:18.972 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 18:07:25 +0000 (0:00:00.036) 0:00:19.009 ********* ok: [/cache/rhel-7.qcow2] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "cryptsetup", "lvm2" ], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 18:07:26 +0000 (0:00:00.910) 0:00:19.919 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 18:07:26 +0000 (0:00:00.043) 0:00:19.963 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 18:07:26 +0000 (0:00:00.031) 0:00:19.995 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 18:07:26 +0000 (0:00:00.034) 0:00:20.030 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 18:07:26 +0000 (0:00:00.031) 0:00:20.061 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "changes": { "installed": [ "cryptsetup" ] }, "rc": 0, "results": [ "7:lvm2-2.02.187-6.el7_9.5.x86_64 providing lvm2 is already installed", "Loaded plugins: search-disabled-repos\nResolving Dependencies\n--> Running transaction check\n---> Package cryptsetup.x86_64 0:2.0.3-6.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package Arch Version Repository Size\n================================================================================\nInstalling:\n cryptsetup x86_64 2.0.3-6.el7 rhel 154 k\n\nTransaction Summary\n================================================================================\nInstall 1 Package\n\nTotal download size: 154 k\nInstalled size: 354 k\nDownloading packages:\nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n Installing : cryptsetup-2.0.3-6.el7.x86_64 1/1 \n Verifying : cryptsetup-2.0.3-6.el7.x86_64 1/1 \n\nInstalled:\n cryptsetup.x86_64 0:2.0.3-6.el7 \n\nComplete!\n" ] } TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 18:07:27 +0000 (0:00:01.470) 0:00:21.532 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 18:07:27 +0000 (0:00:00.029) 0:00:21.562 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 18:07:27 +0000 (0:00:00.082) 0:00:21.644 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 18:07:27 +0000 (0:00:00.021) 0:00:21.666 ********* fatal: [/cache/rhel-7.qcow2]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: encrypted pool 'foo' missing key/password TASK [fedora.linux_system_roles.storage : failed message] ********************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:86 Thursday 21 July 2022 18:07:28 +0000 (0:00:00.876) 0:00:22.543 ********* fatal: [/cache/rhel-7.qcow2]: FAILED! => { "changed": false } MSG: {'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': True, 'pools': [{'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': True, 'encryption_key_size': None, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'fs_type': None, 'mount_options': None, 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': None, 'raid_level': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'deduplication': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'raid_disks': [], 'cache_mode': None, 'name': 'test1', 'cached': None, 'type': None, 'thin': False, 'cache_size': None, 'cache_devices': [], 'fs_create_options': None}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}], 'volumes': [], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': [], 'msg': "encrypted pool 'foo' missing key/password", '_ansible_no_log': False} TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 18:07:28 +0000 (0:00:00.093) 0:00:22.637 ********* TASK [Check that we failed in the role] **************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:53 Thursday 21 July 2022 18:07:28 +0000 (0:00:00.022) 0:00:22.659 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the output of the keyless luks test] ****************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:59 Thursday 21 July 2022 18:07:28 +0000 (0:00:00.036) 0:00:22.696 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Mark tasks to be skipped] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:64 Thursday 21 July 2022 18:07:28 +0000 (0:00:00.049) 0:00:22.746 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_skip_checks": [ "blivet_available", "packages_installed", "service_facts" ] }, "changed": false } TASK [Create a key file] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:73 Thursday 21 July 2022 18:07:28 +0000 (0:00:00.032) 0:00:22.778 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "gid": 0, "group": "root", "mode": "0600", "owner": "root", "path": "/tmp/storage_test4zZk1Mlukskey", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 0, "state": "file", "uid": 0 } TASK [Write the key into the key file] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:80 Thursday 21 July 2022 18:07:29 +0000 (0:00:00.390) 0:00:23.169 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "checksum": "7a4dff3752e2baf5617c57eaac048e2b95e8af91", "dest": "/tmp/storage_test4zZk1Mlukskey", "gid": 0, "group": "root", "md5sum": "4ac07b967150835c00d0865161e48744", "mode": "0600", "owner": "root", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 32, "src": "/root/.ansible/tmp/ansible-tmp-1658426849.3630838-170417-130676408256788/source", "state": "file", "uid": 0 } TASK [Create an encrypted lvm pool using a key file] *************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:87 Thursday 21 July 2022 18:07:29 +0000 (0:00:00.650) 0:00:23.820 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.038) 0:00:23.858 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.035) 0:00:23.893 ********* ok: [/cache/rhel-7.qcow2] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.400) 0:00:24.294 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.062) 0:00:24.357 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.036) 0:00:24.394 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.034) 0:00:24.428 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.056) 0:00:24.485 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.022) 0:00:24.507 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.025) 0:00:24.533 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_key": "/tmp/storage_test4zZk1Mlukskey", "name": "foo", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.035) 0:00:24.569 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.032) 0:00:24.602 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.027) 0:00:24.629 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.024) 0:00:24.653 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.024) 0:00:24.678 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.024) 0:00:24.702 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.056) 0:00:24.758 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 18:07:30 +0000 (0:00:00.019) 0:00:24.778 ********* changed: [/cache/rhel-7.qcow2] => { "actions": [ { "action": "create format", "device": "/dev/nvme1n1", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/nvme1n1", "name": "luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3", "password": "/tmp/storage_test4zZk1Mlukskey", "state": "present" } ], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "cryptsetup", "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": "/tmp/storage_test4zZk1Mlukskey", "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 18:07:38 +0000 (0:00:07.726) 0:00:32.504 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 18:07:38 +0000 (0:00:00.037) 0:00:32.542 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 18:07:38 +0000 (0:00:00.023) 0:00:32.566 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/nvme1n1", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/nvme1n1", "name": "luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3", "password": "/tmp/storage_test4zZk1Mlukskey", "state": "present" } ], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "cryptsetup", "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": "/tmp/storage_test4zZk1Mlukskey", "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 18:07:38 +0000 (0:00:00.040) 0:00:32.606 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": "/tmp/storage_test4zZk1Mlukskey", "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 18:07:38 +0000 (0:00:00.038) 0:00:32.645 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 18:07:38 +0000 (0:00:00.038) 0:00:32.683 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 18:07:38 +0000 (0:00:00.041) 0:00:32.724 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 18:07:39 +0000 (0:00:00.732) 0:00:33.457 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 18:07:40 +0000 (0:00:00.480) 0:00:33.937 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 18:07:40 +0000 (0:00:00.439) 0:00:34.377 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658426836.3731558, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658201031.524, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 70, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658200515.884, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744071677828413", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 18:07:40 +0000 (0:00:00.307) 0:00:34.684 ********* changed: [/cache/rhel-7.qcow2] => (item={'state': 'present', 'password': '/tmp/storage_test4zZk1Mlukskey', 'name': 'luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3', 'backing_device': '/dev/nvme1n1'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/nvme1n1", "name": "luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3", "password": "/tmp/storage_test4zZk1Mlukskey", "state": "present" } } MSG: line added TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 18:07:41 +0000 (0:00:00.467) 0:00:35.151 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:102 Thursday 21 July 2022 18:07:42 +0000 (0:00:00.852) 0:00:36.003 ********* included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 18:07:42 +0000 (0:00:00.039) 0:00:36.043 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_pools_list": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": "/tmp/storage_test4zZk1Mlukskey", "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 18:07:42 +0000 (0:00:00.083) 0:00:36.126 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 18:07:42 +0000 (0:00:00.068) 0:00:36.195 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "4G", "type": "lvm", "uuid": "aad574ee-e377-4a0a-9e7f-7f328c423e99" }, "/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3": { "fstype": "LVM2_member", "label": "", "name": "/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3", "size": "10G", "type": "crypt", "uuid": "i8YNUL-HAYx-bsWb-BQvZ-Utpz-Wh2S-3Qwzmb" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "crypto_LUKS", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "48cdde14-13ef-4118-bd6e-7d993cb73bb3" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-18-06-53-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 18:07:42 +0000 (0:00:00.410) 0:00:36.605 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003531", "end": "2022-07-21 14:07:42.676868", "rc": 0, "start": "2022-07-21 14:07:42.673337" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 18:07:43 +0000 (0:00:00.400) 0:00:37.005 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003379", "end": "2022-07-21 14:07:42.987910", "failed_when_result": false, "rc": 0, "start": "2022-07-21 14:07:42.984531" } STDOUT: luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3 /dev/nvme1n1 /tmp/storage_test4zZk1Mlukskey TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 18:07:43 +0000 (0:00:00.309) 0:00:37.315 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': True, 'encryption_key_size': None, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': '/tmp/storage_test4zZk1Mlukskey', 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 18:07:43 +0000 (0:00:00.058) 0:00:37.374 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 18:07:43 +0000 (0:00:00.033) 0:00:37.407 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members) included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 18:07:43 +0000 (0:00:00.044) 0:00:37.452 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 18:07:43 +0000 (0:00:00.050) 0:00:37.503 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3", "pv": "/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.433) 0:00:37.936 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.047) 0:00:37.983 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.049) 0:00:38.033 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.048) 0:00:38.082 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "crypt" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.042) 0:00:38.125 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.025) 0:00:38.150 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.023) 0:00:38.174 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.046) 0:00:38.220 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml for /cache/rhel-7.qcow2 TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.045) 0:00:38.266 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.024) 0:00:38.291 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.025) 0:00:38.316 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.059) 0:00:38.376 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.023) 0:00:38.400 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.022) 0:00:38.422 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.023) 0:00:38.446 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.023) 0:00:38.469 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.032) 0:00:38.501 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.043) 0:00:38.545 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.044) 0:00:38.590 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.027) 0:00:38.617 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.025) 0:00:38.642 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.027) 0:00:38.669 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml for /cache/rhel-7.qcow2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.041) 0:00:38.711 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.040) 0:00:38.752 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.020) 0:00:38.773 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.020) 0:00:38.794 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.024) 0:00:38.818 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 18:07:44 +0000 (0:00:00.023) 0:00:38.842 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2 TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 18:07:45 +0000 (0:00:00.046) 0:00:38.888 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "1", "_storage_test_expected_crypttab_key_file": "/tmp/storage_test4zZk1Mlukskey" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 18:07:45 +0000 (0:00:00.050) 0:00:38.938 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml for /cache/rhel-7.qcow2 => (item=/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3) TASK [Get the backing device path] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:1 Thursday 21 July 2022 18:07:45 +0000 (0:00:00.041) 0:00:38.979 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "cmd": [ "realpath", "/dev/disk/by-uuid/48cdde14-13ef-4118-bd6e-7d993cb73bb3" ], "delta": "0:00:00.004660", "end": "2022-07-21 14:07:44.973527", "rc": 0, "start": "2022-07-21 14:07:44.968867" } STDOUT: /dev/nvme1n1 TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:6 Thursday 21 July 2022 18:07:45 +0000 (0:00:00.323) 0:00:39.303 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this member] *************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:11 Thursday 21 July 2022 18:07:46 +0000 (0:00:00.557) 0:00:39.860 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cryptsetup", "luksDump", "/dev/nvme1n1" ], "delta": "0:00:00.112840", "end": "2022-07-21 14:07:45.955279", "rc": 0, "start": "2022-07-21 14:07:45.842439" } STDOUT: LUKS header information for /dev/nvme1n1 Version: 1 Cipher name: aes Cipher mode: xts-plain64 Hash spec: sha256 Payload offset: 4096 MK bits: 512 MK digest: 0f 79 40 53 be 58 93 82 ac cb 28 77 f8 f9 72 1e 1d 91 80 e7 MK salt: a9 da ea 3a 85 d5 d6 53 48 a5 ed 4a 4b 4d 61 f1 23 c2 b2 8b ac f0 8a 8f b9 f0 a3 1d 8e b7 f3 ac MK iterations: 22787 UUID: 48cdde14-13ef-4118-bd6e-7d993cb73bb3 Key Slot 0: ENABLED Iterations: 364594 Salt: 7c e0 d3 0e 6e e6 59 c2 df a0 c4 44 61 ad 58 e6 86 6c ac c7 84 e5 78 03 ff 75 6b 01 c2 d8 66 e5 Key material offset: 8 AF stripes: 4000 Key Slot 1: DISABLED Key Slot 2: DISABLED Key Slot 3: DISABLED Key Slot 4: DISABLED Key Slot 5: DISABLED Key Slot 6: DISABLED Key Slot 7: DISABLED TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:17 Thursday 21 July 2022 18:07:46 +0000 (0:00:00.431) 0:00:40.292 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:23 Thursday 21 July 2022 18:07:46 +0000 (0:00:00.025) 0:00:40.318 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:29 Thursday 21 July 2022 18:07:46 +0000 (0:00:00.025) 0:00:40.343 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 18:07:46 +0000 (0:00:00.026) 0:00:40.369 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 18:07:46 +0000 (0:00:00.046) 0:00:40.416 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [ "luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3 /dev/nvme1n1 /tmp/storage_test4zZk1Mlukskey" ] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 18:07:46 +0000 (0:00:00.052) 0:00:40.468 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 18:07:46 +0000 (0:00:00.050) 0:00:40.519 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 18:07:46 +0000 (0:00:00.050) 0:00:40.569 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 18:07:46 +0000 (0:00:00.035) 0:00:40.605 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 18:07:46 +0000 (0:00:00.054) 0:00:40.659 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 18:07:46 +0000 (0:00:00.034) 0:00:40.693 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 18:07:46 +0000 (0:00:00.034) 0:00:40.728 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 18:07:46 +0000 (0:00:00.048) 0:00:40.776 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 18:07:46 +0000 (0:00:00.045) 0:00:40.822 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 18:07:46 +0000 (0:00:00.026) 0:00:40.848 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.024) 0:00:40.873 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.024) 0:00:40.897 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.024) 0:00:40.922 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.026) 0:00:40.948 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.024) 0:00:40.973 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.024) 0:00:40.997 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.036) 0:00:41.034 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.038) 0:00:41.072 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.042) 0:00:41.114 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.050) 0:00:41.164 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.083) 0:00:41.248 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.048) 0:00:41.297 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1037768, "block_size": 4096, "block_total": 1046016, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4250697728, "size_total": 4284481536, "uuid": "aad574ee-e377-4a0a-9e7f-7f328c423e99" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1037768, "block_size": 4096, "block_total": 1046016, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4250697728, "size_total": 4284481536, "uuid": "aad574ee-e377-4a0a-9e7f-7f328c423e99" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.064) 0:00:41.361 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.051) 0:00:41.412 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.084) 0:00:41.497 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.084) 0:00:41.582 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.056) 0:00:41.638 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.023) 0:00:41.662 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.022) 0:00:41.684 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.036) 0:00:41.721 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.057) 0:00:41.779 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:07:47 +0000 (0:00:00.047) 0:00:41.826 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:07:48 +0000 (0:00:00.055) 0:00:41.881 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:07:48 +0000 (0:00:00.039) 0:00:41.921 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:07:48 +0000 (0:00:00.035) 0:00:41.956 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:07:48 +0000 (0:00:00.041) 0:00:41.997 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:07:48 +0000 (0:00:00.051) 0:00:42.049 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658426858.1291556, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658426858.1291556, "dev": 5, "device_type": 64513, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 35875, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658426858.1291556, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:07:48 +0000 (0:00:00.318) 0:00:42.368 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:07:48 +0000 (0:00:00.039) 0:00:42.407 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:07:48 +0000 (0:00:00.039) 0:00:42.447 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:07:48 +0000 (0:00:00.038) 0:00:42.486 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:07:48 +0000 (0:00:00.026) 0:00:42.512 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:07:48 +0000 (0:00:00.036) 0:00:42.548 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:07:48 +0000 (0:00:00.023) 0:00:42.572 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.522) 0:00:43.094 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.025) 0:00:43.120 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.024) 0:00:43.145 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.055) 0:00:43.200 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.023) 0:00:43.223 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.022) 0:00:43.245 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.023) 0:00:43.269 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.022) 0:00:43.292 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.023) 0:00:43.315 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.050) 0:00:43.366 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.046) 0:00:43.413 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.033) 0:00:43.446 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.032) 0:00:43.479 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.033) 0:00:43.512 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.038) 0:00:43.551 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.035) 0:00:43.587 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.035) 0:00:43.622 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.033) 0:00:43.656 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.036) 0:00:43.692 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.039) 0:00:43.731 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.034) 0:00:43.766 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:07:49 +0000 (0:00:00.035) 0:00:43.801 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:07:50 +0000 (0:00:00.457) 0:00:44.258 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:07:50 +0000 (0:00:00.360) 0:00:44.618 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "4294967296" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:07:50 +0000 (0:00:00.051) 0:00:44.669 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:07:50 +0000 (0:00:00.071) 0:00:44.741 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:07:50 +0000 (0:00:00.035) 0:00:44.777 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.094) 0:00:44.871 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.035) 0:00:44.906 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.036) 0:00:44.943 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.038) 0:00:44.981 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.034) 0:00:45.015 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.037) 0:00:45.053 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.050) 0:00:45.103 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.024157", "end": "2022-07-21 14:07:51.108432", "rc": 0, "start": "2022-07-21 14:07:51.084275" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.332) 0:00:45.436 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.051) 0:00:45.487 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.051) 0:00:45.539 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.039) 0:00:45.578 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.037) 0:00:45.616 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.037) 0:00:45.653 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.035) 0:00:45.689 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.036) 0:00:45.725 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.021) 0:00:45.747 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Remove the key file] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:104 Thursday 21 July 2022 18:07:51 +0000 (0:00:00.032) 0:00:45.779 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "path": "/tmp/storage_test4zZk1Mlukskey", "state": "absent" } TASK [create a file] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/create-test-file.yml:10 Thursday 21 July 2022 18:07:52 +0000 (0:00:00.434) 0:00:46.214 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "dest": "/opt/test1/quux", "gid": 0, "group": "root", "mode": "0644", "owner": "root", "secontext": "unconfined_u:object_r:unlabeled_t:s0", "size": 0, "state": "file", "uid": 0 } TASK [Remove the encryption layer] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:114 Thursday 21 July 2022 18:07:52 +0000 (0:00:00.317) 0:00:46.532 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 18:07:52 +0000 (0:00:00.039) 0:00:46.572 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 18:07:52 +0000 (0:00:00.035) 0:00:46.607 ********* ok: [/cache/rhel-7.qcow2] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 18:07:53 +0000 (0:00:00.405) 0:00:47.013 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 18:07:53 +0000 (0:00:00.089) 0:00:47.102 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 18:07:53 +0000 (0:00:00.033) 0:00:47.136 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 18:07:53 +0000 (0:00:00.092) 0:00:47.229 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 18:07:53 +0000 (0:00:00.112) 0:00:47.342 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 18:07:53 +0000 (0:00:00.020) 0:00:47.362 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 18:07:53 +0000 (0:00:00.027) 0:00:47.390 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "nvme1n1" ], "encryption": false, "encryption_password": "yabbadabbadoo", "name": "foo", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 18:07:53 +0000 (0:00:00.040) 0:00:47.431 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 18:07:53 +0000 (0:00:00.037) 0:00:47.468 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 18:07:53 +0000 (0:00:00.028) 0:00:47.497 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 18:07:53 +0000 (0:00:00.028) 0:00:47.525 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 18:07:53 +0000 (0:00:00.034) 0:00:47.560 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 18:07:53 +0000 (0:00:00.029) 0:00:47.590 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 18:07:53 +0000 (0:00:00.058) 0:00:47.648 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 18:07:53 +0000 (0:00:00.023) 0:00:47.671 ********* fatal: [/cache/rhel-7.qcow2]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: cannot remove and recreate existing pool 'foo' in safe mode TASK [fedora.linux_system_roles.storage : failed message] ********************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:86 Thursday 21 July 2022 18:07:54 +0000 (0:00:01.156) 0:00:48.828 ********* fatal: [/cache/rhel-7.qcow2]: FAILED! => { "changed": false } MSG: {'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': True, 'pools': [{'name': 'foo', 'encryption_password': 'yabbadabbadoo', 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': 0, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': 'luks1', 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'fs_type': None, 'mount_options': None, 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': None, 'raid_level': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'deduplication': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'raid_disks': [], 'cache_mode': None, 'name': 'test1', 'cached': None, 'type': None, 'thin': False, 'cache_size': None, 'cache_devices': [], 'fs_create_options': None}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}], 'volumes': [], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': [], 'msg': "cannot remove and recreate existing pool 'foo' in safe mode", '_ansible_no_log': False} TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 18:07:55 +0000 (0:00:00.041) 0:00:48.870 ********* TASK [Check that we failed in the role] **************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:132 Thursday 21 July 2022 18:07:55 +0000 (0:00:00.021) 0:00:48.892 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the output of the safe_mode test] ********************************* task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:138 Thursday 21 July 2022 18:07:55 +0000 (0:00:00.035) 0:00:48.927 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [stat the file] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-data-preservation.yml:10 Thursday 21 July 2022 18:07:55 +0000 (0:00:00.052) 0:00:48.980 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658426872.2021558, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658426872.2021558, "dev": 64513, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 67, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0644", "mtime": 1658426872.2021558, "nlink": 1, "path": "/opt/test1/quux", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "version": "1921240464", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [assert file presence] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-data-preservation.yml:15 Thursday 21 July 2022 18:07:55 +0000 (0:00:00.294) 0:00:49.274 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Remove the encryption layer] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:145 Thursday 21 July 2022 18:07:55 +0000 (0:00:00.041) 0:00:49.316 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 18:07:55 +0000 (0:00:00.042) 0:00:49.359 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 18:07:55 +0000 (0:00:00.035) 0:00:49.394 ********* ok: [/cache/rhel-7.qcow2] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 18:07:55 +0000 (0:00:00.428) 0:00:49.823 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 18:07:56 +0000 (0:00:00.062) 0:00:49.885 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 18:07:56 +0000 (0:00:00.065) 0:00:49.951 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 18:07:56 +0000 (0:00:00.069) 0:00:50.020 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 18:07:56 +0000 (0:00:00.055) 0:00:50.075 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 18:07:56 +0000 (0:00:00.031) 0:00:50.107 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 18:07:56 +0000 (0:00:00.074) 0:00:50.181 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "nvme1n1" ], "encryption": false, "encryption_password": "yabbadabbadoo", "name": "foo", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 18:07:56 +0000 (0:00:00.037) 0:00:50.219 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 18:07:56 +0000 (0:00:00.038) 0:00:50.257 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 18:07:56 +0000 (0:00:00.029) 0:00:50.287 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 18:07:56 +0000 (0:00:00.026) 0:00:50.314 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 18:07:56 +0000 (0:00:00.027) 0:00:50.341 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 18:07:56 +0000 (0:00:00.025) 0:00:50.367 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 18:07:56 +0000 (0:00:00.056) 0:00:50.424 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 18:07:56 +0000 (0:00:00.021) 0:00:50.446 ********* changed: [/cache/rhel-7.qcow2] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3", "fs_type": null }, { "action": "destroy format", "device": "/dev/nvme1n1", "fs_type": "luks" }, { "action": "create format", "device": "/dev/nvme1n1", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/nvme1n1", "name": "luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3", "password": "-", "state": "absent" } ], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "nvme1n1" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 18:07:58 +0000 (0:00:02.226) 0:00:52.672 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 18:07:58 +0000 (0:00:00.039) 0:00:52.712 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 18:07:58 +0000 (0:00:00.022) 0:00:52.734 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/mapper/luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3", "fs_type": null }, { "action": "destroy format", "device": "/dev/nvme1n1", "fs_type": "luks" }, { "action": "create format", "device": "/dev/nvme1n1", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/nvme1n1", "name": "luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3", "password": "-", "state": "absent" } ], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "nvme1n1" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 18:07:58 +0000 (0:00:00.050) 0:00:52.785 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "nvme1n1" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 18:07:58 +0000 (0:00:00.038) 0:00:52.823 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 18:07:59 +0000 (0:00:00.038) 0:00:52.862 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 18:07:59 +0000 (0:00:00.336) 0:00:53.199 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 18:07:59 +0000 (0:00:00.483) 0:00:53.682 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 18:08:00 +0000 (0:00:00.355) 0:00:54.038 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 18:08:00 +0000 (0:00:00.477) 0:00:54.515 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658426862.9871557, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "b98d06a3373b8c7e8e0dd11c155590cbe7299f42", "ctime": 1658426860.8161557, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 8596711, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0600", "mtime": 1658426860.8151557, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 86, "uid": 0, "version": "1360316211", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 18:08:00 +0000 (0:00:00.317) 0:00:54.832 ********* changed: [/cache/rhel-7.qcow2] => (item={'state': 'absent', 'password': '-', 'name': 'luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3', 'backing_device': '/dev/nvme1n1'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/nvme1n1", "name": "luks-48cdde14-13ef-4118-bd6e-7d993cb73bb3", "password": "-", "state": "absent" }, "found": 1 } MSG: 1 line(s) removed TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 18:08:01 +0000 (0:00:00.353) 0:00:55.186 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:161 Thursday 21 July 2022 18:08:02 +0000 (0:00:00.992) 0:00:56.178 ********* included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 18:08:02 +0000 (0:00:00.037) 0:00:56.216 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_pools_list": [ { "disks": [ "nvme1n1" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 18:08:02 +0000 (0:00:00.054) 0:00:56.271 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 18:08:02 +0000 (0:00:00.040) 0:00:56.311 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "4G", "type": "lvm", "uuid": "338e3995-c397-48c5-91bd-00d3b7b12d1b" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "LVM2_member", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "VqSiGV-M1tJ-jLWC-Faab-HOlY-Jiyf-hcOI3d" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-18-06-53-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 18:08:02 +0000 (0:00:00.330) 0:00:56.642 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003486", "end": "2022-07-21 14:08:02.605326", "rc": 0, "start": "2022-07-21 14:08:02.601840" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 18:08:03 +0000 (0:00:00.289) 0:00:56.932 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003409", "end": "2022-07-21 14:08:02.908874", "failed_when_result": false, "rc": 0, "start": "2022-07-21 14:08:02.905465" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 18:08:03 +0000 (0:00:00.307) 0:00:57.239 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': 'yabbadabbadoo', 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': 0, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': 'luks1', 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 18:08:03 +0000 (0:00:00.057) 0:00:57.296 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 18:08:03 +0000 (0:00:00.031) 0:00:57.327 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members) included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 18:08:03 +0000 (0:00:00.044) 0:00:57.372 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/nvme1n1" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 18:08:03 +0000 (0:00:00.050) 0:00:57.423 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/nvme1n1", "pv": "/dev/nvme1n1" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 18:08:03 +0000 (0:00:00.310) 0:00:57.734 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 18:08:03 +0000 (0:00:00.051) 0:00:57.785 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/nvme1n1" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 18:08:03 +0000 (0:00:00.051) 0:00:57.836 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.049) 0:00:57.886 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.035) 0:00:57.922 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.048) 0:00:57.971 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.022) 0:00:57.993 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/nvme1n1" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.040) 0:00:58.034 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml for /cache/rhel-7.qcow2 TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.040) 0:00:58.074 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.024) 0:00:58.098 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.022) 0:00:58.121 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.023) 0:00:58.144 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.047) 0:00:58.191 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.023) 0:00:58.215 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.023) 0:00:58.239 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.023) 0:00:58.262 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.037) 0:00:58.299 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.041) 0:00:58.341 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.043) 0:00:58.384 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.028) 0:00:58.412 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.028) 0:00:58.440 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.030) 0:00:58.471 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml for /cache/rhel-7.qcow2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.044) 0:00:58.515 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.043) 0:00:58.558 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.023) 0:00:58.582 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.022) 0:00:58.605 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.024) 0:00:58.630 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.025) 0:00:58.655 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2 TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.046) 0:00:58.702 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.049) 0:00:58.752 ********* skipping: [/cache/rhel-7.qcow2] => (item=/dev/nvme1n1) => { "_storage_test_pool_member_path": "/dev/nvme1n1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.029) 0:00:58.781 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/nvme1n1) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 18:08:04 +0000 (0:00:00.043) 0:00:58.825 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.052) 0:00:58.877 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.046) 0:00:58.923 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.034) 0:00:58.958 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.036) 0:00:58.995 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.037) 0:00:59.033 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.031) 0:00:59.065 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.035) 0:00:59.100 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.045) 0:00:59.146 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.041) 0:00:59.188 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.022) 0:00:59.210 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.024) 0:00:59.235 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.024) 0:00:59.259 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.023) 0:00:59.282 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.023) 0:00:59.306 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.050) 0:00:59.356 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.023) 0:00:59.379 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.033) 0:00:59.413 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.036) 0:00:59.450 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.039) 0:00:59.490 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.045) 0:00:59.536 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.077) 0:00:59.614 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.037) 0:00:59.652 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1037768, "block_size": 4096, "block_total": 1046016, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4250697728, "size_total": 4284481536, "uuid": "338e3995-c397-48c5-91bd-00d3b7b12d1b" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1037768, "block_size": 4096, "block_total": 1046016, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4250697728, "size_total": 4284481536, "uuid": "338e3995-c397-48c5-91bd-00d3b7b12d1b" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.050) 0:00:59.703 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.046) 0:00:59.749 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.045) 0:00:59.795 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:08:05 +0000 (0:00:00.047) 0:00:59.842 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.021) 0:00:59.864 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.021) 0:00:59.885 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.022) 0:00:59.908 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.031) 0:00:59.940 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.058) 0:00:59.998 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.047) 0:01:00.045 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.046) 0:01:00.092 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.033) 0:01:00.125 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.032) 0:01:00.158 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.037) 0:01:00.195 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.037) 0:01:00.232 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658426878.3001556, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658426878.3001556, "dev": 5, "device_type": 64512, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 41404, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658426878.3001556, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.305) 0:01:00.537 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.035) 0:01:00.573 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.036) 0:01:00.610 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.034) 0:01:00.644 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.024) 0:01:00.668 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.044) 0:01:00.712 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:08:06 +0000 (0:00:00.023) 0:01:00.736 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.534) 0:01:01.270 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.024) 0:01:01.295 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.058) 0:01:01.353 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.059) 0:01:01.413 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.022) 0:01:01.436 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.022) 0:01:01.458 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.021) 0:01:01.480 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.021) 0:01:01.501 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.021) 0:01:01.523 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.049) 0:01:01.572 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.045) 0:01:01.618 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.033) 0:01:01.651 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.033) 0:01:01.684 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.036) 0:01:01.720 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.031) 0:01:01.751 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.034) 0:01:01.786 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:08:07 +0000 (0:00:00.033) 0:01:01.820 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:08:08 +0000 (0:00:00.037) 0:01:01.857 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:08:08 +0000 (0:00:00.033) 0:01:01.890 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:08:08 +0000 (0:00:00.033) 0:01:01.923 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:08:08 +0000 (0:00:00.033) 0:01:01.957 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:08:08 +0000 (0:00:00.041) 0:01:01.998 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:08:08 +0000 (0:00:00.290) 0:01:02.288 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:08:08 +0000 (0:00:00.306) 0:01:02.595 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "4294967296" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:08:08 +0000 (0:00:00.048) 0:01:02.644 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:08:08 +0000 (0:00:00.034) 0:01:02.679 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:08:08 +0000 (0:00:00.036) 0:01:02.715 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:08:08 +0000 (0:00:00.034) 0:01:02.750 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:08:08 +0000 (0:00:00.035) 0:01:02.785 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:08:08 +0000 (0:00:00.034) 0:01:02.820 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:08:09 +0000 (0:00:00.036) 0:01:02.857 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:08:09 +0000 (0:00:00.033) 0:01:02.890 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:08:09 +0000 (0:00:00.034) 0:01:02.924 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:08:09 +0000 (0:00:00.050) 0:01:02.975 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.026050", "end": "2022-07-21 14:08:08.964484", "rc": 0, "start": "2022-07-21 14:08:08.938434" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:08:09 +0000 (0:00:00.319) 0:01:03.294 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:08:09 +0000 (0:00:00.081) 0:01:03.376 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:08:09 +0000 (0:00:00.084) 0:01:03.461 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:08:09 +0000 (0:00:00.039) 0:01:03.501 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:08:09 +0000 (0:00:00.036) 0:01:03.537 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:08:09 +0000 (0:00:00.038) 0:01:03.576 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:08:09 +0000 (0:00:00.039) 0:01:03.615 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 18:08:09 +0000 (0:00:00.065) 0:01:03.680 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 18:08:09 +0000 (0:00:00.021) 0:01:03.702 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [create a file] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/create-test-file.yml:10 Thursday 21 July 2022 18:08:09 +0000 (0:00:00.032) 0:01:03.734 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "dest": "/opt/test1/quux", "gid": 0, "group": "root", "mode": "0644", "owner": "root", "secontext": "unconfined_u:object_r:unlabeled_t:s0", "size": 0, "state": "file", "uid": 0 } TASK [Add encryption to the pool] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:167 Thursday 21 July 2022 18:08:10 +0000 (0:00:00.303) 0:01:04.038 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 18:08:10 +0000 (0:00:00.037) 0:01:04.075 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 18:08:10 +0000 (0:00:00.033) 0:01:04.109 ********* ok: [/cache/rhel-7.qcow2] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 18:08:10 +0000 (0:00:00.387) 0:01:04.497 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 18:08:10 +0000 (0:00:00.058) 0:01:04.555 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 18:08:10 +0000 (0:00:00.032) 0:01:04.588 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 18:08:10 +0000 (0:00:00.031) 0:01:04.619 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 18:08:10 +0000 (0:00:00.052) 0:01:04.672 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 18:08:10 +0000 (0:00:00.019) 0:01:04.692 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 18:08:10 +0000 (0:00:00.024) 0:01:04.717 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": "serpent-xts-plain64", "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 18:08:10 +0000 (0:00:00.036) 0:01:04.753 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 18:08:10 +0000 (0:00:00.036) 0:01:04.789 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 18:08:10 +0000 (0:00:00.034) 0:01:04.823 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 18:08:10 +0000 (0:00:00.027) 0:01:04.850 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 18:08:11 +0000 (0:00:00.025) 0:01:04.876 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 18:08:11 +0000 (0:00:00.024) 0:01:04.900 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 18:08:11 +0000 (0:00:00.052) 0:01:04.953 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 18:08:11 +0000 (0:00:00.020) 0:01:04.974 ********* fatal: [/cache/rhel-7.qcow2]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: cannot remove and recreate existing pool 'foo' in safe mode TASK [fedora.linux_system_roles.storage : failed message] ********************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:86 Thursday 21 July 2022 18:08:12 +0000 (0:00:01.078) 0:01:06.052 ********* fatal: [/cache/rhel-7.qcow2]: FAILED! => { "changed": false } MSG: {'crypts': [], 'mounts': [], 'leaves': [], 'changed': False, 'actions': [], 'failed': True, 'volumes': [], 'invocation': {'module_args': {'packages_only': False, 'disklabel_type': None, 'diskvolume_mkfs_option_map': {'ext4': '-F', 'ext3': '-F', 'ext2': '-F'}, 'safe_mode': True, 'pools': [{'name': 'foo', 'encryption_password': 'yabbadabbadoo', 'state': 'present', 'raid_metadata_version': None, 'encryption': True, 'encryption_key_size': 512, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': 'luks1', 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'fs_type': None, 'mount_options': None, 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, 'encryption': None, 'raid_level': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'deduplication': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'raid_disks': [], 'cache_mode': None, 'name': 'test1', 'cached': None, 'type': None, 'thin': False, 'cache_size': None, 'cache_devices': [], 'fs_create_options': None}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': 'serpent-xts-plain64', 'raid_spare_count': None}], 'volumes': [], 'pool_defaults': {'encryption_password': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_cipher': None, 'disks': [], 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_device_count': None, 'state': 'present', 'volumes': [], 'raid_chunk_size': None, 'type': 'lvm', 'raid_level': None, 'raid_spare_count': None}, 'volume_defaults': {'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'xfs', 'mount_options': 'defaults', 'size': 0, 'mount_point': '', 'compression': None, 'encryption_password': None, 'encryption': False, 'raid_level': None, 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'fs_overwrite_existing': True, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'mount_passno': 0, 'raid_spare_count': None, 'cache_mode': None, 'deduplication': None, 'cached': False, 'type': 'lvm', 'disks': [], 'thin_pool_size': None, 'thin': None, 'mount_check': 0, 'cache_size': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, 'use_partitions': None}}, 'pools': [], 'packages': [], 'msg': "cannot remove and recreate existing pool 'foo' in safe mode", '_ansible_no_log': False} TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 18:08:12 +0000 (0:00:00.040) 0:01:06.092 ********* TASK [Check that we failed in the role] **************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:188 Thursday 21 July 2022 18:08:12 +0000 (0:00:00.022) 0:01:06.115 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the output of the safe_mode test] ********************************* task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:194 Thursday 21 July 2022 18:08:12 +0000 (0:00:00.036) 0:01:06.152 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [stat the file] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-data-preservation.yml:10 Thursday 21 July 2022 18:08:12 +0000 (0:00:00.050) 0:01:06.202 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658426889.7101557, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658426889.7101557, "dev": 64512, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 67, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0644", "mtime": 1658426889.7101557, "nlink": 1, "path": "/opt/test1/quux", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "version": "18446744073650502484", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [assert file presence] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-data-preservation.yml:15 Thursday 21 July 2022 18:08:12 +0000 (0:00:00.319) 0:01:06.522 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Add encryption to the pool] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:201 Thursday 21 July 2022 18:08:12 +0000 (0:00:00.071) 0:01:06.594 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 18:08:12 +0000 (0:00:00.042) 0:01:06.636 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 18:08:12 +0000 (0:00:00.035) 0:01:06.672 ********* ok: [/cache/rhel-7.qcow2] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 18:08:13 +0000 (0:00:00.453) 0:01:07.126 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 18:08:13 +0000 (0:00:00.066) 0:01:07.193 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 18:08:13 +0000 (0:00:00.037) 0:01:07.231 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 18:08:13 +0000 (0:00:00.037) 0:01:07.269 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 18:08:13 +0000 (0:00:00.055) 0:01:07.324 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 18:08:13 +0000 (0:00:00.022) 0:01:07.346 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 18:08:13 +0000 (0:00:00.028) 0:01:07.375 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": "serpent-xts-plain64", "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 18:08:13 +0000 (0:00:00.037) 0:01:07.413 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 18:08:13 +0000 (0:00:00.038) 0:01:07.452 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 18:08:13 +0000 (0:00:00.027) 0:01:07.479 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 18:08:13 +0000 (0:00:00.026) 0:01:07.506 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 18:08:13 +0000 (0:00:00.029) 0:01:07.535 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 18:08:13 +0000 (0:00:00.029) 0:01:07.565 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 18:08:13 +0000 (0:00:00.059) 0:01:07.625 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 18:08:13 +0000 (0:00:00.020) 0:01:07.646 ********* changed: [/cache/rhel-7.qcow2] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/nvme1n1", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/nvme1n1", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/nvme1n1", "name": "luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "password": "-", "state": "present" } ], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "cryptsetup", "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": "serpent-xts-plain64", "encryption_key": null, "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 18:08:22 +0000 (0:00:08.217) 0:01:15.863 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 18:08:22 +0000 (0:00:00.035) 0:01:15.898 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 18:08:22 +0000 (0:00:00.020) 0:01:15.919 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/nvme1n1", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/nvme1n1", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/nvme1n1", "name": "luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "password": "-", "state": "present" } ], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "cryptsetup", "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": "serpent-xts-plain64", "encryption_key": null, "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 18:08:22 +0000 (0:00:00.039) 0:01:15.959 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": "serpent-xts-plain64", "encryption_key": null, "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 18:08:22 +0000 (0:00:00.037) 0:01:15.996 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 18:08:22 +0000 (0:00:00.040) 0:01:16.036 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 18:08:22 +0000 (0:00:00.351) 0:01:16.388 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 18:08:23 +0000 (0:00:00.486) 0:01:16.875 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 18:08:23 +0000 (0:00:00.415) 0:01:17.290 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 18:08:23 +0000 (0:00:00.433) 0:01:17.724 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658426882.9081557, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658426880.8501558, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 4211252, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658426880.8501558, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1286599433", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 18:08:24 +0000 (0:00:00.355) 0:01:18.079 ********* changed: [/cache/rhel-7.qcow2] => (item={'state': 'present', 'password': '-', 'name': 'luks-63f0ca91-2836-4052-b2bd-bad50b2f763a', 'backing_device': '/dev/nvme1n1'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/nvme1n1", "name": "luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "password": "-", "state": "present" } } MSG: line added TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 18:08:24 +0000 (0:00:00.341) 0:01:18.421 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:220 Thursday 21 July 2022 18:08:25 +0000 (0:00:00.858) 0:01:19.279 ********* included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 18:08:25 +0000 (0:00:00.038) 0:01:19.318 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_pools_list": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": "serpent-xts-plain64", "encryption_key": null, "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 18:08:25 +0000 (0:00:00.050) 0:01:19.368 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 18:08:25 +0000 (0:00:00.036) 0:01:19.405 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "4G", "type": "lvm", "uuid": "c1ae1d42-6141-452f-a8e0-39fe28e7a39f" }, "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a": { "fstype": "LVM2_member", "label": "", "name": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "size": "10G", "type": "crypt", "uuid": "1p0vOA-WQHg-Nbi2-fXwC-b8xc-6Dr3-PTZgux" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "crypto_LUKS", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "63f0ca91-2836-4052-b2bd-bad50b2f763a" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-18-06-53-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 18:08:25 +0000 (0:00:00.321) 0:01:19.727 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003659", "end": "2022-07-21 14:08:25.711735", "rc": 0, "start": "2022-07-21 14:08:25.708076" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 18:08:26 +0000 (0:00:00.316) 0:01:20.044 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003258", "end": "2022-07-21 14:08:26.020841", "failed_when_result": false, "rc": 0, "start": "2022-07-21 14:08:26.017583" } STDOUT: luks-63f0ca91-2836-4052-b2bd-bad50b2f763a /dev/nvme1n1 - TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 18:08:26 +0000 (0:00:00.309) 0:01:20.353 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': 'yabbadabbadoo', 'state': 'present', 'raid_metadata_version': None, 'encryption': True, 'encryption_key_size': 512, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': 'luks1', 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': 'serpent-xts-plain64', 'raid_spare_count': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 18:08:26 +0000 (0:00:00.059) 0:01:20.412 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 18:08:26 +0000 (0:00:00.035) 0:01:20.448 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members) included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 18:08:26 +0000 (0:00:00.047) 0:01:20.496 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 18:08:26 +0000 (0:00:00.056) 0:01:20.552 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "pv": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.315) 0:01:20.867 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.080) 0:01:20.947 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.081) 0:01:21.029 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.047) 0:01:21.077 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "crypt" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.065) 0:01:21.143 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.023) 0:01:21.166 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.025) 0:01:21.191 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.039) 0:01:21.231 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml for /cache/rhel-7.qcow2 TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.040) 0:01:21.272 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.052) 0:01:21.324 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.026) 0:01:21.350 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.024) 0:01:21.375 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.024) 0:01:21.400 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.026) 0:01:21.426 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.023) 0:01:21.450 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.025) 0:01:21.475 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.036) 0:01:21.511 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.045) 0:01:21.557 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.042) 0:01:21.599 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.026) 0:01:21.626 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.029) 0:01:21.655 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.029) 0:01:21.685 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml for /cache/rhel-7.qcow2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.044) 0:01:21.729 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.046) 0:01:21.775 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.023) 0:01:21.799 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.023) 0:01:21.822 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 18:08:27 +0000 (0:00:00.022) 0:01:21.844 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 18:08:28 +0000 (0:00:00.022) 0:01:21.867 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2 TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 18:08:28 +0000 (0:00:00.044) 0:01:21.911 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "1", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 18:08:28 +0000 (0:00:00.047) 0:01:21.959 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml for /cache/rhel-7.qcow2 => (item=/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a) TASK [Get the backing device path] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:1 Thursday 21 July 2022 18:08:28 +0000 (0:00:00.045) 0:01:22.004 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "cmd": [ "realpath", "/dev/disk/by-uuid/63f0ca91-2836-4052-b2bd-bad50b2f763a" ], "delta": "0:00:00.003560", "end": "2022-07-21 14:08:28.004919", "rc": 0, "start": "2022-07-21 14:08:28.001359" } STDOUT: /dev/nvme1n1 TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:6 Thursday 21 July 2022 18:08:28 +0000 (0:00:00.334) 0:01:22.339 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this member] *************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:11 Thursday 21 July 2022 18:08:29 +0000 (0:00:00.548) 0:01:22.887 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cryptsetup", "luksDump", "/dev/nvme1n1" ], "delta": "0:00:00.050581", "end": "2022-07-21 14:08:28.974441", "rc": 0, "start": "2022-07-21 14:08:28.923860" } STDOUT: LUKS header information for /dev/nvme1n1 Version: 1 Cipher name: serpent Cipher mode: xts-plain64 Hash spec: sha256 Payload offset: 4096 MK bits: 512 MK digest: 70 e6 51 f2 03 51 04 e6 b6 82 a4 17 a4 80 02 98 25 45 77 af MK salt: 73 ac 94 06 ec 76 49 d8 47 c8 e9 04 54 78 69 e4 31 48 14 99 69 69 8e 10 17 15 f5 89 2c 65 f0 06 MK iterations: 22914 UUID: 63f0ca91-2836-4052-b2bd-bad50b2f763a Key Slot 0: ENABLED Iterations: 365102 Salt: 77 d4 46 06 16 29 89 89 18 da 89 5a 84 dc 58 b2 06 78 c4 2f bc 0f 34 3f 26 ff 54 23 05 20 bc da Key material offset: 8 AF stripes: 4000 Key Slot 1: DISABLED Key Slot 2: DISABLED Key Slot 3: DISABLED Key Slot 4: DISABLED Key Slot 5: DISABLED Key Slot 6: DISABLED Key Slot 7: DISABLED TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:17 Thursday 21 July 2022 18:08:29 +0000 (0:00:00.428) 0:01:23.316 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:23 Thursday 21 July 2022 18:08:29 +0000 (0:00:00.131) 0:01:23.447 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:29 Thursday 21 July 2022 18:08:29 +0000 (0:00:00.054) 0:01:23.502 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 18:08:29 +0000 (0:00:00.052) 0:01:23.554 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 18:08:29 +0000 (0:00:00.048) 0:01:23.602 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [ "luks-63f0ca91-2836-4052-b2bd-bad50b2f763a /dev/nvme1n1 -" ] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 18:08:29 +0000 (0:00:00.052) 0:01:23.655 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 18:08:29 +0000 (0:00:00.049) 0:01:23.705 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 18:08:29 +0000 (0:00:00.051) 0:01:23.757 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 18:08:29 +0000 (0:00:00.035) 0:01:23.792 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 18:08:29 +0000 (0:00:00.049) 0:01:23.841 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.038) 0:01:23.879 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.034) 0:01:23.914 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.048) 0:01:23.963 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.046) 0:01:24.009 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.025) 0:01:24.035 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.026) 0:01:24.062 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.024) 0:01:24.086 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.024) 0:01:24.111 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.025) 0:01:24.137 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.025) 0:01:24.162 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.024) 0:01:24.187 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.032) 0:01:24.219 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.034) 0:01:24.254 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '4g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.043) 0:01:24.298 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.051) 0:01:24.349 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.081) 0:01:24.430 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.042) 0:01:24.473 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1037768, "block_size": 4096, "block_total": 1046016, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4250697728, "size_total": 4284481536, "uuid": "c1ae1d42-6141-452f-a8e0-39fe28e7a39f" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1037768, "block_size": 4096, "block_total": 1046016, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4250697728, "size_total": 4284481536, "uuid": "c1ae1d42-6141-452f-a8e0-39fe28e7a39f" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.056) 0:01:24.529 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.052) 0:01:24.582 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.051) 0:01:24.633 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.050) 0:01:24.684 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.024) 0:01:24.709 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.024) 0:01:24.734 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.026) 0:01:24.760 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:08:30 +0000 (0:00:00.078) 0:01:24.839 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:08:31 +0000 (0:00:00.063) 0:01:24.902 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:08:31 +0000 (0:00:00.052) 0:01:24.954 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:08:31 +0000 (0:00:00.052) 0:01:25.007 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:08:31 +0000 (0:00:00.038) 0:01:25.045 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:08:31 +0000 (0:00:00.036) 0:01:25.081 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:08:31 +0000 (0:00:00.043) 0:01:25.125 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:08:31 +0000 (0:00:00.039) 0:01:25.164 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658426901.4921558, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658426901.4921558, "dev": 5, "device_type": 64513, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 50251, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658426901.4921558, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:08:31 +0000 (0:00:00.327) 0:01:25.491 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:08:31 +0000 (0:00:00.037) 0:01:25.529 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:08:31 +0000 (0:00:00.039) 0:01:25.568 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:08:31 +0000 (0:00:00.033) 0:01:25.602 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:08:31 +0000 (0:00:00.023) 0:01:25.625 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:08:31 +0000 (0:00:00.038) 0:01:25.664 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:08:31 +0000 (0:00:00.026) 0:01:25.690 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.554) 0:01:26.244 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.024) 0:01:26.269 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.023) 0:01:26.292 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.051) 0:01:26.344 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.024) 0:01:26.368 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.022) 0:01:26.391 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.023) 0:01:26.414 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.023) 0:01:26.437 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.025) 0:01:26.462 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.051) 0:01:26.514 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.047) 0:01:26.562 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.038) 0:01:26.600 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.037) 0:01:26.638 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.037) 0:01:26.676 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.075) 0:01:26.752 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.039) 0:01:26.792 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:08:32 +0000 (0:00:00.036) 0:01:26.828 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:08:33 +0000 (0:00:00.037) 0:01:26.865 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:08:33 +0000 (0:00:00.038) 0:01:26.904 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:08:33 +0000 (0:00:00.103) 0:01:27.007 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:08:33 +0000 (0:00:00.038) 0:01:27.046 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:08:33 +0000 (0:00:00.038) 0:01:27.085 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:08:33 +0000 (0:00:00.328) 0:01:27.414 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:08:33 +0000 (0:00:00.331) 0:01:27.745 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "4294967296" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:08:33 +0000 (0:00:00.055) 0:01:27.800 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:08:33 +0000 (0:00:00.040) 0:01:27.841 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.041) 0:01:27.882 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.039) 0:01:27.922 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.038) 0:01:27.961 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.038) 0:01:28.000 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.042) 0:01:28.042 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.045) 0:01:28.088 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.041) 0:01:28.129 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.058) 0:01:28.187 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.026035", "end": "2022-07-21 14:08:34.198577", "rc": 0, "start": "2022-07-21 14:08:34.172542" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.343) 0:01:28.531 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.054) 0:01:28.585 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.055) 0:01:28.641 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.042) 0:01:28.683 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.036) 0:01:28.720 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.035) 0:01:28.755 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.035) 0:01:28.791 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.035) 0:01:28.826 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 18:08:34 +0000 (0:00:00.020) 0:01:28.847 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [create a file] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/create-test-file.yml:10 Thursday 21 July 2022 18:08:35 +0000 (0:00:00.033) 0:01:28.880 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "dest": "/opt/test1/quux", "gid": 0, "group": "root", "mode": "0644", "owner": "root", "secontext": "unconfined_u:object_r:unlabeled_t:s0", "size": 0, "state": "file", "uid": 0 } TASK [Change the mountpoint, leaving encryption in place] ********************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:224 Thursday 21 July 2022 18:08:35 +0000 (0:00:00.316) 0:01:29.197 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 18:08:35 +0000 (0:00:00.047) 0:01:29.245 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 18:08:35 +0000 (0:00:00.034) 0:01:29.279 ********* ok: [/cache/rhel-7.qcow2] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 18:08:35 +0000 (0:00:00.408) 0:01:29.688 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 18:08:35 +0000 (0:00:00.100) 0:01:29.789 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 18:08:36 +0000 (0:00:00.111) 0:01:29.900 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 18:08:36 +0000 (0:00:00.038) 0:01:29.939 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 18:08:36 +0000 (0:00:00.059) 0:01:29.998 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 18:08:36 +0000 (0:00:00.022) 0:01:30.021 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 18:08:36 +0000 (0:00:00.030) 0:01:30.051 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "nvme1n1" ], "name": "foo", "type": "lvm", "volumes": [ { "mount_point": "/opt/test2", "name": "test1" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 18:08:36 +0000 (0:00:00.038) 0:01:30.090 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 18:08:36 +0000 (0:00:00.038) 0:01:30.129 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 18:08:36 +0000 (0:00:00.027) 0:01:30.157 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 18:08:36 +0000 (0:00:00.026) 0:01:30.183 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 18:08:36 +0000 (0:00:00.031) 0:01:30.215 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 18:08:36 +0000 (0:00:00.034) 0:01:30.250 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 18:08:36 +0000 (0:00:00.079) 0:01:30.329 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 18:08:36 +0000 (0:00:00.023) 0:01:30.352 ********* ok: [/cache/rhel-7.qcow2] => { "actions": [], "changed": false, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1" ], "mounts": [ { "path": "/opt/test1", "state": "absent" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "cryptsetup", "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "nvme1n1" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 4294967296, "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 18:08:37 +0000 (0:00:01.262) 0:01:31.615 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 18:08:37 +0000 (0:00:00.038) 0:01:31.653 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 18:08:37 +0000 (0:00:00.021) 0:01:31.675 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1" ], "mounts": [ { "path": "/opt/test1", "state": "absent" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "cryptsetup", "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "nvme1n1" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 4294967296, "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 18:08:37 +0000 (0:00:00.041) 0:01:31.716 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "nvme1n1" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 4294967296, "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 18:08:37 +0000 (0:00:00.042) 0:01:31.759 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 18:08:37 +0000 (0:00:00.039) 0:01:31.798 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'path': '/opt/test1', 'state': 'absent'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "mount_info": { "path": "/opt/test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 18:08:38 +0000 (0:00:00.363) 0:01:32.162 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 18:08:38 +0000 (0:00:00.465) 0:01:32.627 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test2', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 18:08:39 +0000 (0:00:00.378) 0:01:33.006 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 18:08:39 +0000 (0:00:00.459) 0:01:33.466 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658426906.0201557, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "1ae0b225b3f52246a053e66fb0fd684f653e9452", "ctime": 1658426904.0811558, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 16777316, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0600", "mtime": 1658426904.0801556, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 57, "uid": 0, "version": "1251431158", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 18:08:39 +0000 (0:00:00.332) 0:01:33.799 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 18:08:39 +0000 (0:00:00.028) 0:01:33.828 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [Assert to implicitly preserve encryption on existing pool] *************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:236 Thursday 21 July 2022 18:08:40 +0000 (0:00:00.841) 0:01:34.670 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [stat the file] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-data-preservation.yml:10 Thursday 21 July 2022 18:08:40 +0000 (0:00:00.039) 0:01:34.709 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658426914.8611557, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658426914.8611557, "dev": 64513, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 67, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0644", "mtime": 1658426914.8611557, "nlink": 1, "path": "/opt/test2/quux", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "version": "18446744071811286532", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [assert file presence] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-data-preservation.yml:15 Thursday 21 July 2022 18:08:41 +0000 (0:00:00.365) 0:01:35.075 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:245 Thursday 21 July 2022 18:08:41 +0000 (0:00:00.114) 0:01:35.189 ********* included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 18:08:41 +0000 (0:00:00.041) 0:01:35.231 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_pools_list": [ { "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "nvme1n1" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 4294967296, "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 18:08:41 +0000 (0:00:00.053) 0:01:35.284 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 18:08:41 +0000 (0:00:00.037) 0:01:35.322 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "4G", "type": "lvm", "uuid": "c1ae1d42-6141-452f-a8e0-39fe28e7a39f" }, "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a": { "fstype": "LVM2_member", "label": "", "name": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "size": "10G", "type": "crypt", "uuid": "1p0vOA-WQHg-Nbi2-fXwC-b8xc-6Dr3-PTZgux" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "crypto_LUKS", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "63f0ca91-2836-4052-b2bd-bad50b2f763a" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-18-06-53-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 18:08:41 +0000 (0:00:00.315) 0:01:35.638 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003425", "end": "2022-07-21 14:08:41.613346", "rc": 0, "start": "2022-07-21 14:08:41.609921" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 /dev/mapper/foo-test1 /opt/test2 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 18:08:42 +0000 (0:00:00.310) 0:01:35.948 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003767", "end": "2022-07-21 14:08:41.935598", "failed_when_result": false, "rc": 0, "start": "2022-07-21 14:08:41.931831" } STDOUT: luks-63f0ca91-2836-4052-b2bd-bad50b2f763a /dev/nvme1n1 - TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 18:08:42 +0000 (0:00:00.320) 0:01:36.269 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': True, 'encryption_key_size': 0, 'disks': ['nvme1n1'], 'raid_level': None, 'encryption_luks_version': 'luks1', 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': 4294967296, 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 18:08:42 +0000 (0:00:00.062) 0:01:36.331 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 18:08:42 +0000 (0:00:00.035) 0:01:36.367 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members) included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 18:08:42 +0000 (0:00:00.047) 0:01:36.415 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 18:08:42 +0000 (0:00:00.061) 0:01:36.477 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "pv": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 18:08:42 +0000 (0:00:00.331) 0:01:36.808 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.053) 0:01:36.862 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.051) 0:01:36.913 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.051) 0:01:36.965 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "crypt" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.039) 0:01:37.004 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.024) 0:01:37.029 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.022) 0:01:37.051 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.040) 0:01:37.092 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml for /cache/rhel-7.qcow2 TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.041) 0:01:37.133 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.023) 0:01:37.156 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.023) 0:01:37.180 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.058) 0:01:37.239 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.024) 0:01:37.263 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.024) 0:01:37.287 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.024) 0:01:37.312 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.028) 0:01:37.340 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.034) 0:01:37.375 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.044) 0:01:37.419 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': 4294967296, 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.043) 0:01:37.462 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.028) 0:01:37.490 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.028) 0:01:37.519 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.029) 0:01:37.549 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml for /cache/rhel-7.qcow2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.043) 0:01:37.592 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': 4294967296, 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.046) 0:01:37.638 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.024) 0:01:37.663 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.022) 0:01:37.686 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.024) 0:01:37.710 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.026) 0:01:37.737 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2 TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.047) 0:01:37.784 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "1", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 18:08:43 +0000 (0:00:00.052) 0:01:37.837 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml for /cache/rhel-7.qcow2 => (item=/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a) TASK [Get the backing device path] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:1 Thursday 21 July 2022 18:08:44 +0000 (0:00:00.044) 0:01:37.881 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "cmd": [ "realpath", "/dev/disk/by-uuid/63f0ca91-2836-4052-b2bd-bad50b2f763a" ], "delta": "0:00:00.003663", "end": "2022-07-21 14:08:43.881652", "rc": 0, "start": "2022-07-21 14:08:43.877989" } STDOUT: /dev/nvme1n1 TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:6 Thursday 21 July 2022 18:08:44 +0000 (0:00:00.329) 0:01:38.211 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this member] *************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:11 Thursday 21 July 2022 18:08:44 +0000 (0:00:00.532) 0:01:38.743 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cryptsetup", "luksDump", "/dev/nvme1n1" ], "delta": "0:00:00.035890", "end": "2022-07-21 14:08:44.777769", "rc": 0, "start": "2022-07-21 14:08:44.741879" } STDOUT: LUKS header information for /dev/nvme1n1 Version: 1 Cipher name: serpent Cipher mode: xts-plain64 Hash spec: sha256 Payload offset: 4096 MK bits: 512 MK digest: 70 e6 51 f2 03 51 04 e6 b6 82 a4 17 a4 80 02 98 25 45 77 af MK salt: 73 ac 94 06 ec 76 49 d8 47 c8 e9 04 54 78 69 e4 31 48 14 99 69 69 8e 10 17 15 f5 89 2c 65 f0 06 MK iterations: 22914 UUID: 63f0ca91-2836-4052-b2bd-bad50b2f763a Key Slot 0: ENABLED Iterations: 365102 Salt: 77 d4 46 06 16 29 89 89 18 da 89 5a 84 dc 58 b2 06 78 c4 2f bc 0f 34 3f 26 ff 54 23 05 20 bc da Key material offset: 8 AF stripes: 4000 Key Slot 1: DISABLED Key Slot 2: DISABLED Key Slot 3: DISABLED Key Slot 4: DISABLED Key Slot 5: DISABLED Key Slot 6: DISABLED Key Slot 7: DISABLED TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:17 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.369) 0:01:39.113 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:23 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.108) 0:01:39.222 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-encryption.yml:29 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.026) 0:01:39.248 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.026) 0:01:39.274 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.081) 0:01:39.356 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [ "luks-63f0ca91-2836-4052-b2bd-bad50b2f763a /dev/nvme1n1 -" ] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.052) 0:01:39.408 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.049) 0:01:39.457 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.049) 0:01:39.507 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.040) 0:01:39.548 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.059) 0:01:39.608 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.034) 0:01:39.642 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.033) 0:01:39.675 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.046) 0:01:39.722 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': 4294967296, 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.043) 0:01:39.766 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.023) 0:01:39.789 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.023) 0:01:39.813 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 18:08:45 +0000 (0:00:00.023) 0:01:39.837 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.022) 0:01:39.860 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.022) 0:01:39.882 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.023) 0:01:39.906 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.025) 0:01:39.931 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.037) 0:01:39.969 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.038) 0:01:40.008 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': 4294967296, 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.044) 0:01:40.052 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.049) 0:01:40.101 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.080) 0:01:40.181 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.046) 0:01:40.228 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1037768, "block_size": 4096, "block_total": 1046016, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097148, "inode_total": 2097152, "inode_used": 4, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4250697728, "size_total": 4284481536, "uuid": "c1ae1d42-6141-452f-a8e0-39fe28e7a39f" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1037768, "block_size": 4096, "block_total": 1046016, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097148, "inode_total": 2097152, "inode_used": 4, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4250697728, "size_total": 4284481536, "uuid": "c1ae1d42-6141-452f-a8e0-39fe28e7a39f" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.052) 0:01:40.281 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.049) 0:01:40.330 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.052) 0:01:40.383 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.052) 0:01:40.436 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.026) 0:01:40.462 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.024) 0:01:40.487 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.023) 0:01:40.510 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.032) 0:01:40.543 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.100) 0:01:40.644 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.129) 0:01:40.773 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:08:46 +0000 (0:00:00.051) 0:01:40.825 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:08:47 +0000 (0:00:00.038) 0:01:40.863 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:08:47 +0000 (0:00:00.039) 0:01:40.903 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:08:47 +0000 (0:00:00.040) 0:01:40.943 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:08:47 +0000 (0:00:00.040) 0:01:40.984 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658426901.4921558, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658426901.4921558, "dev": 5, "device_type": 64513, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 50251, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658426901.4921558, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:08:47 +0000 (0:00:00.326) 0:01:41.311 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:08:47 +0000 (0:00:00.037) 0:01:41.348 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:08:47 +0000 (0:00:00.037) 0:01:41.386 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:08:47 +0000 (0:00:00.039) 0:01:41.425 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:08:47 +0000 (0:00:00.024) 0:01:41.449 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:08:47 +0000 (0:00:00.043) 0:01:41.492 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:08:47 +0000 (0:00:00.023) 0:01:41.515 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.525) 0:01:42.041 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.026) 0:01:42.067 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.024) 0:01:42.092 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.054) 0:01:42.146 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.025) 0:01:42.172 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.026) 0:01:42.198 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.025) 0:01:42.223 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.026) 0:01:42.250 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.022) 0:01:42.273 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.051) 0:01:42.324 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.049) 0:01:42.374 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.039) 0:01:42.413 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.035) 0:01:42.448 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.033) 0:01:42.482 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.036) 0:01:42.518 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.039) 0:01:42.557 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.036) 0:01:42.594 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.038) 0:01:42.633 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.037) 0:01:42.671 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.040) 0:01:42.711 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.037) 0:01:42.749 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:08:48 +0000 (0:00:00.037) 0:01:42.786 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:08:49 +0000 (0:00:00.362) 0:01:43.149 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:08:49 +0000 (0:00:00.328) 0:01:43.477 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "4294967296" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:08:49 +0000 (0:00:00.088) 0:01:43.565 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:08:49 +0000 (0:00:00.114) 0:01:43.680 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:08:49 +0000 (0:00:00.037) 0:01:43.718 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:08:49 +0000 (0:00:00.035) 0:01:43.753 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:08:49 +0000 (0:00:00.034) 0:01:43.788 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:08:49 +0000 (0:00:00.036) 0:01:43.824 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:08:50 +0000 (0:00:00.034) 0:01:43.859 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:08:50 +0000 (0:00:00.034) 0:01:43.894 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:08:50 +0000 (0:00:00.034) 0:01:43.929 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:08:50 +0000 (0:00:00.051) 0:01:43.980 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.024141", "end": "2022-07-21 14:08:49.985817", "rc": 0, "start": "2022-07-21 14:08:49.961676" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:08:50 +0000 (0:00:00.342) 0:01:44.323 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:08:50 +0000 (0:00:00.050) 0:01:44.373 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:08:50 +0000 (0:00:00.052) 0:01:44.426 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:08:50 +0000 (0:00:00.038) 0:01:44.465 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:08:50 +0000 (0:00:00.038) 0:01:44.504 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:08:50 +0000 (0:00:00.039) 0:01:44.543 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:08:50 +0000 (0:00:00.040) 0:01:44.584 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 18:08:50 +0000 (0:00:00.033) 0:01:44.617 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 18:08:50 +0000 (0:00:00.022) 0:01:44.639 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Clean up] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:247 Thursday 21 July 2022 18:08:50 +0000 (0:00:00.038) 0:01:44.678 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 18:08:50 +0000 (0:00:00.054) 0:01:44.733 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 18:08:50 +0000 (0:00:00.039) 0:01:44.772 ********* ok: [/cache/rhel-7.qcow2] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 18:08:51 +0000 (0:00:00.418) 0:01:45.191 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 18:08:51 +0000 (0:00:00.067) 0:01:45.258 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 18:08:51 +0000 (0:00:00.035) 0:01:45.294 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 18:08:51 +0000 (0:00:00.036) 0:01:45.331 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 18:08:51 +0000 (0:00:00.056) 0:01:45.387 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 18:08:51 +0000 (0:00:00.021) 0:01:45.409 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 18:08:51 +0000 (0:00:00.030) 0:01:45.439 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 18:08:51 +0000 (0:00:00.118) 0:01:45.558 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": [ { "disks": [ "nvme1n1" ], "name": "foo", "state": "absent", "type": "disk" } ] } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 18:08:51 +0000 (0:00:00.038) 0:01:45.597 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 18:08:51 +0000 (0:00:00.031) 0:01:45.628 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 18:08:51 +0000 (0:00:00.029) 0:01:45.657 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 18:08:51 +0000 (0:00:00.031) 0:01:45.689 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 18:08:51 +0000 (0:00:00.029) 0:01:45.718 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 18:08:51 +0000 (0:00:00.059) 0:01:45.778 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 18:08:51 +0000 (0:00:00.021) 0:01:45.800 ********* changed: [/cache/rhel-7.qcow2] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "fs_type": null }, { "action": "destroy format", "device": "/dev/nvme1n1", "fs_type": "luks" } ], "changed": true, "crypts": [ { "backing_device": "/dev/nvme1n1", "name": "luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "password": "-", "state": "absent" } ], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "xfsprogs" ], "pools": [], "volumes": [ { "_device": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "_mount_id": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "_raw_device": "/dev/nvme1n1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "lvmpv", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 10735321088, "state": "absent", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 18:08:53 +0000 (0:00:01.777) 0:01:47.577 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 18:08:53 +0000 (0:00:00.039) 0:01:47.617 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 18:08:53 +0000 (0:00:00.022) 0:01:47.639 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "fs_type": null }, { "action": "destroy format", "device": "/dev/nvme1n1", "fs_type": "luks" } ], "changed": true, "crypts": [ { "backing_device": "/dev/nvme1n1", "name": "luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "password": "-", "state": "absent" } ], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "xfsprogs" ], "pools": [], "volumes": [ { "_device": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "_mount_id": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "_raw_device": "/dev/nvme1n1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "lvmpv", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 10735321088, "state": "absent", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 18:08:53 +0000 (0:00:00.046) 0:01:47.686 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 18:08:53 +0000 (0:00:00.039) 0:01:47.725 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [ { "_device": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "_mount_id": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "_raw_device": "/dev/nvme1n1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "lvmpv", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 10735321088, "state": "absent", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 18:08:53 +0000 (0:00:00.042) 0:01:47.768 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'state': 'absent', 'path': '/opt/test2', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 18:08:54 +0000 (0:00:00.354) 0:01:48.123 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 18:08:54 +0000 (0:00:00.455) 0:01:48.578 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 18:08:54 +0000 (0:00:00.037) 0:01:48.615 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 18:08:55 +0000 (0:00:00.423) 0:01:49.039 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658426906.0201557, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "1ae0b225b3f52246a053e66fb0fd684f653e9452", "ctime": 1658426904.0811558, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 16777316, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0600", "mtime": 1658426904.0801556, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 57, "uid": 0, "version": "1251431158", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 18:08:55 +0000 (0:00:00.298) 0:01:49.337 ********* changed: [/cache/rhel-7.qcow2] => (item={'state': 'absent', 'password': '-', 'name': 'luks-63f0ca91-2836-4052-b2bd-bad50b2f763a', 'backing_device': '/dev/nvme1n1'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/nvme1n1", "name": "luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "password": "-", "state": "absent" }, "found": 1 } MSG: 1 line(s) removed TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 18:08:55 +0000 (0:00:00.335) 0:01:49.672 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_luks_pool.yml:257 Thursday 21 July 2022 18:08:56 +0000 (0:00:00.837) 0:01:50.510 ********* included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 18:08:56 +0000 (0:00:00.046) 0:01:50.557 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Print out volume information] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 18:08:56 +0000 (0:00:00.038) 0:01:50.595 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_volumes_list": [ { "_device": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "_mount_id": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a", "_raw_device": "/dev/nvme1n1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "nvme1n1" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "lvmpv", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 10735321088, "state": "absent", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 18:08:56 +0000 (0:00:00.094) 0:01:50.689 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-18-06-53-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 18:08:57 +0000 (0:00:00.355) 0:01:51.045 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.004135", "end": "2022-07-21 14:08:57.027693", "rc": 0, "start": "2022-07-21 14:08:57.023558" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 18:08:57 +0000 (0:00:00.316) 0:01:51.361 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003925", "end": "2022-07-21 14:08:57.341794", "failed_when_result": false, "rc": 0, "start": "2022-07-21 14:08:57.337869" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 18:08:57 +0000 (0:00:00.315) 0:01:51.677 ********* TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 18:08:57 +0000 (0:00:00.020) 0:01:51.698 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/nvme1n1', 'raid_metadata_version': None, 'mount_device_identifier': 'uuid', 'fs_type': 'lvmpv', 'mount_options': 'defaults', '_device': '/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a', 'size': 10735321088, 'mount_point': None, 'compression': None, 'encryption_password': None, 'encryption': True, 'raid_level': None, 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'type': 'disk', 'encryption_key_size': 0, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': 'luks1', 'cache_size': 0, '_mount_id': '/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a', 'raid_spare_count': None, 'name': 'foo', 'cache_mode': None, 'cache_devices': [], 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['nvme1n1'], 'thin': None, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'thin_pool_size': None, 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:08:57 +0000 (0:00:00.056) 0:01:51.754 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:08:57 +0000 (0:00:00.047) 0:01:51.801 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.075) 0:01:51.876 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/luks-63f0ca91-2836-4052-b2bd-bad50b2f763a" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.040) 0:01:51.916 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.056) 0:01:51.973 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.025) 0:01:51.998 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.050) 0:01:52.048 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.037) 0:01:52.086 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.023) 0:01:52.109 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.021) 0:01:52.131 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.022) 0:01:52.153 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.037) 0:01:52.190 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.060) 0:01:52.251 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.021) 0:01:52.272 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.050) 0:01:52.323 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.036) 0:01:52.360 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.035) 0:01:52.395 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.024) 0:01:52.420 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.023) 0:01:52.444 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658426933.1981556, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658426933.1981556, "dev": 5, "device_type": 66306, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 9588, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/blockdevice", "mode": "0660", "mtime": 1658426933.1981556, "nlink": 1, "path": "/dev/nvme1n1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.307) 0:01:52.751 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:08:58 +0000 (0:00:00.069) 0:01:52.821 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:08:59 +0000 (0:00:00.049) 0:01:52.870 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "disk" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:08:59 +0000 (0:00:00.037) 0:01:52.907 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:08:59 +0000 (0:00:00.024) 0:01:52.932 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:08:59 +0000 (0:00:00.043) 0:01:52.975 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "exists": false } } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:08:59 +0000 (0:00:00.302) 0:01:53.278 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:08:59 +0000 (0:00:00.550) 0:01:53.829 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:08:59 +0000 (0:00:00.025) 0:01:53.855 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.039) 0:01:53.895 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.024) 0:01:53.920 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.024) 0:01:53.944 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.022) 0:01:53.966 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.021) 0:01:53.987 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.020) 0:01:54.008 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.022) 0:01:54.031 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.053) 0:01:54.084 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.045) 0:01:54.130 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.032) 0:01:54.163 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.034) 0:01:54.198 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.032) 0:01:54.230 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.030) 0:01:54.261 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.032) 0:01:54.294 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.040) 0:01:54.334 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.036) 0:01:54.370 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.034) 0:01:54.405 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.035) 0:01:54.441 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.037) 0:01:54.478 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.036) 0:01:54.515 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.022) 0:01:54.537 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.034) 0:01:54.571 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.037) 0:01:54.609 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.035) 0:01:54.644 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.034) 0:01:54.679 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.037) 0:01:54.716 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.043) 0:01:54.759 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.037) 0:01:54.797 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:09:00 +0000 (0:00:00.035) 0:01:54.833 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:09:01 +0000 (0:00:00.091) 0:01:54.924 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "4294967296" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:09:01 +0000 (0:00:00.037) 0:01:54.962 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:09:01 +0000 (0:00:00.022) 0:01:54.984 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:09:01 +0000 (0:00:00.022) 0:01:55.007 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:09:01 +0000 (0:00:00.022) 0:01:55.030 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:09:01 +0000 (0:00:00.063) 0:01:55.093 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:09:01 +0000 (0:00:00.024) 0:01:55.117 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:09:01 +0000 (0:00:00.022) 0:01:55.140 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:09:01 +0000 (0:00:00.022) 0:01:55.163 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:09:01 +0000 (0:00:00.023) 0:01:55.186 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 18:09:01 +0000 (0:00:00.031) 0:01:55.218 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* /cache/rhel-7.qcow2 : ok=543 changed=25 unreachable=0 failed=3 skipped=371 rescued=3 ignored=0 Thursday 21 July 2022 18:09:01 +0000 (0:00:00.073) 0:01:55.292 ********* =============================================================================== fedora.linux_system_roles.storage : make sure blivet is available ------- 9.00s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 8.22s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 7.73s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 2.23s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.78s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : make sure required packages are installed --- 1.47s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 set up internal repositories -------------------------------------------- 1.42s /cache/rhel-7_setup.yml:5 ----------------------------------------------------- fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.26s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.16s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : get service facts ------------------- 1.11s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.08s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Gathering Facts --------------------------------------------------------- 1.00s /tmp/tmptomayb7j/tests/storage/tests_luks_pool_nvme_generated.yml:3 ----------- fedora.linux_system_roles.storage : Update facts ------------------------ 0.99s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 fedora.linux_system_roles.storage : get required packages --------------- 0.91s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 0.88s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : Update facts ------------------------ 0.86s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 fedora.linux_system_roles.storage : Update facts ------------------------ 0.85s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 fedora.linux_system_roles.storage : Update facts ------------------------ 0.84s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 fedora.linux_system_roles.storage : Update facts ------------------------ 0.84s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 fedora.linux_system_roles.storage : Update facts ------------------------ 0.84s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183