ansible-playbook [core 2.12.6] config file = /etc/ansible/ansible.cfg configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.9/site-packages/ansible ansible collection location = /tmp/tmp5bkr4li_ executable location = /usr/bin/ansible-playbook python version = 3.9.13 (main, May 18 2022, 00:00:00) [GCC 11.3.1 20220421 (Red Hat 11.3.1-2)] jinja version = 2.11.3 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: rhel-7_setup.yml ***************************************************** 1 plays in /cache/rhel-7_setup.yml PLAY [Setup repos] ************************************************************* META: ran handlers TASK [set up internal repositories] ******************************************** task path: /cache/rhel-7_setup.yml:5 Thursday 21 July 2022 15:24:08 +0000 (0:00:00.018) 0:00:00.018 ********* changed: [/cache/rhel-7.qcow2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [/cache/rhel-7.qcow2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [/cache/rhel-7.qcow2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [/cache/rhel-7.qcow2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [/cache/rhel-7.qcow2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* /cache/rhel-7.qcow2 : ok=1 changed=1 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 Thursday 21 July 2022 15:24:09 +0000 (0:00:01.414) 0:00:01.432 ********* =============================================================================== set up internal repositories -------------------------------------------- 1.41s /cache/rhel-7_setup.yml:5 ----------------------------------------------------- PLAYBOOK: tests_lvm_one_disk_multiple_volumes_scsi_generated.yml *************** 2 plays in /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes_scsi_generated.yml PLAY [all] ********************************************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes_scsi_generated.yml:3 Thursday 21 July 2022 15:24:09 +0000 (0:00:00.015) 0:00:01.448 ********* ok: [/cache/rhel-7.qcow2] META: ran handlers TASK [set disk interface for test] ********************************************* task path: /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes_scsi_generated.yml:7 Thursday 21 July 2022 15:24:10 +0000 (0:00:01.013) 0:00:02.462 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_use_interface": "scsi" }, "changed": false } META: ran handlers META: ran handlers PLAY [all] ********************************************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes.yml:2 Thursday 21 July 2022 15:24:10 +0000 (0:00:00.044) 0:00:02.507 ********* ok: [/cache/rhel-7.qcow2] META: ran handlers TASK [include_role : linux-system-roles.storage] ******************************* task path: /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes.yml:11 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.754) 0:00:03.261 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.034) 0:00:03.296 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.030) 0:00:03.327 ********* ok: [/cache/rhel-7.qcow2] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.418) 0:00:03.745 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.055) 0:00:03.800 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.030) 0:00:03.831 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.030) 0:00:03.862 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.045) 0:00:03.907 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 15:24:11 +0000 (0:00:00.016) 0:00:03.924 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "changes": { "installed": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "rc": 0, "results": [ "Loaded plugins: search-disabled-repos\nResolving Dependencies\n--> Running transaction check\n---> Package libblockdev-crypto.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libblockdev-utils(x86-64) = 2.18-5.el7 for package: libblockdev-crypto-2.18-5.el7.x86_64\n--> Processing Dependency: libvolume_key.so.1()(64bit) for package: libblockdev-crypto-2.18-5.el7.x86_64\n--> Processing Dependency: libbd_utils.so.2()(64bit) for package: libblockdev-crypto-2.18-5.el7.x86_64\n---> Package libblockdev-dm.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libdmraid.so.1(Base)(64bit) for package: libblockdev-dm-2.18-5.el7.x86_64\n--> Processing Dependency: dmraid for package: libblockdev-dm-2.18-5.el7.x86_64\n--> Processing Dependency: libdmraid.so.1()(64bit) for package: libblockdev-dm-2.18-5.el7.x86_64\n---> Package libblockdev-lvm.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: lvm2 for package: libblockdev-lvm-2.18-5.el7.x86_64\n--> Processing Dependency: device-mapper-persistent-data for package: libblockdev-lvm-2.18-5.el7.x86_64\n---> Package libblockdev-mdraid.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: mdadm for package: libblockdev-mdraid-2.18-5.el7.x86_64\n--> Processing Dependency: libbytesize.so.1()(64bit) for package: libblockdev-mdraid-2.18-5.el7.x86_64\n---> Package libblockdev-swap.x86_64 0:2.18-5.el7 will be installed\n---> Package python-enum34.noarch 0:1.0.4-1.el7 will be installed\n---> Package python2-blivet3.noarch 1:3.1.3-3.el7 will be installed\n--> Processing Dependency: blivet3-data = 1:3.1.3-3.el7 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-bytesize >= 0.3 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-blockdev >= 2.17 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: pyparted >= 3.9 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-hawkey for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: lsof for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Running transaction check\n---> Package blivet3-data.noarch 1:3.1.3-3.el7 will be installed\n---> Package device-mapper-persistent-data.x86_64 0:0.8.5-3.el7_9.2 will be installed\n--> Processing Dependency: libaio.so.1(LIBAIO_0.4)(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n--> Processing Dependency: libaio.so.1(LIBAIO_0.1)(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n--> Processing Dependency: libaio.so.1()(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n---> Package dmraid.x86_64 0:1.0.0.rc16-28.el7 will be installed\n--> Processing Dependency: libdevmapper-event.so.1.02(Base)(64bit) for package: dmraid-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: dmraid-events for package: dmraid-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: libdevmapper-event.so.1.02()(64bit) for package: dmraid-1.0.0.rc16-28.el7.x86_64\n---> Package libblockdev-utils.x86_64 0:2.18-5.el7 will be installed\n---> Package libbytesize.x86_64 0:1.2-1.el7 will be installed\n--> Processing Dependency: libmpfr.so.4()(64bit) for package: libbytesize-1.2-1.el7.x86_64\n---> Package lsof.x86_64 0:4.87-6.el7 will be installed\n---> Package lvm2.x86_64 7:2.02.187-6.el7_9.5 will be installed\n--> Processing Dependency: lvm2-libs = 7:2.02.187-6.el7_9.5 for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n--> Processing Dependency: liblvm2app.so.2.2(Base)(64bit) for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n--> Processing Dependency: liblvm2app.so.2.2()(64bit) for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n---> Package mdadm.x86_64 0:4.1-9.el7_9 will be installed\n--> Processing Dependency: libreport-filesystem for package: mdadm-4.1-9.el7_9.x86_64\n---> Package pyparted.x86_64 1:3.9-15.el7 will be installed\n---> Package python2-blockdev.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libblockdev(x86-64) = 2.18-5.el7 for package: python2-blockdev-2.18-5.el7.x86_64\n---> Package python2-bytesize.x86_64 0:1.2-1.el7 will be installed\n---> Package python2-hawkey.x86_64 0:0.22.5-2.el7_9 will be installed\n--> Processing Dependency: libdnf(x86-64) = 0.22.5-2.el7_9 for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: python2-libdnf = 0.22.5-2.el7_9 for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolv.so.0(SOLV_1.0)(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolvext.so.0(SOLV_1.0)(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libdnf.so.2()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libjson-glib-1.0.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libmodulemd.so.1()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: librepo.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: librhsm.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolv.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolvext.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n---> Package volume_key-libs.x86_64 0:0.3.9-9.el7 will be installed\n--> Running transaction check\n---> Package device-mapper-event-libs.x86_64 7:1.02.170-6.el7_9.5 will be installed\n---> Package dmraid-events.x86_64 0:1.0.0.rc16-28.el7 will be installed\n--> Processing Dependency: sgpio for package: dmraid-events-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: device-mapper-event for package: dmraid-events-1.0.0.rc16-28.el7.x86_64\n---> Package json-glib.x86_64 0:1.4.2-2.el7 will be installed\n---> Package libaio.x86_64 0:0.3.109-13.el7 will be installed\n---> Package libblockdev.x86_64 0:2.18-5.el7 will be installed\n---> Package libdnf.x86_64 0:0.22.5-2.el7_9 will be installed\n---> Package libmodulemd.x86_64 0:1.6.3-1.el7 will be installed\n---> Package librepo.x86_64 0:1.8.1-8.el7_9 will be installed\n---> Package libreport-filesystem.x86_64 0:2.1.11-53.el7 will be installed\n---> Package librhsm.x86_64 0:0.0.3-3.el7_9 will be installed\n---> Package libsolv.x86_64 0:0.6.34-4.el7 will be installed\n---> Package lvm2-libs.x86_64 7:2.02.187-6.el7_9.5 will be installed\n---> Package mpfr.x86_64 0:3.1.1-4.el7 will be installed\n---> Package python2-libdnf.x86_64 0:0.22.5-2.el7_9 will be installed\n--> Running transaction check\n---> Package device-mapper-event.x86_64 7:1.02.170-6.el7_9.5 will be installed\n---> Package sgpio.x86_64 0:1.2.0.10-13.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package Arch Version Repository Size\n================================================================================\nInstalling:\n libblockdev-crypto x86_64 2.18-5.el7 rhel 60 k\n libblockdev-dm x86_64 2.18-5.el7 rhel-optional 54 k\n libblockdev-lvm x86_64 2.18-5.el7 rhel 62 k\n libblockdev-mdraid x86_64 2.18-5.el7 rhel 57 k\n libblockdev-swap x86_64 2.18-5.el7 rhel 52 k\n python-enum34 noarch 1.0.4-1.el7 rhel 52 k\n python2-blivet3 noarch 1:3.1.3-3.el7 rhel 851 k\nInstalling for dependencies:\n blivet3-data noarch 1:3.1.3-3.el7 rhel 77 k\n device-mapper-event\n x86_64 7:1.02.170-6.el7_9.5 rhel 192 k\n device-mapper-event-libs\n x86_64 7:1.02.170-6.el7_9.5 rhel 192 k\n device-mapper-persistent-data\n x86_64 0.8.5-3.el7_9.2 rhel 423 k\n dmraid x86_64 1.0.0.rc16-28.el7 rhel 151 k\n dmraid-events x86_64 1.0.0.rc16-28.el7 rhel 21 k\n json-glib x86_64 1.4.2-2.el7 rhel 134 k\n libaio x86_64 0.3.109-13.el7 rhel 24 k\n libblockdev x86_64 2.18-5.el7 rhel 119 k\n libblockdev-utils x86_64 2.18-5.el7 rhel 59 k\n libbytesize x86_64 1.2-1.el7 rhel 52 k\n libdnf x86_64 0.22.5-2.el7_9 rhel-7-server-extras-rpms 536 k\n libmodulemd x86_64 1.6.3-1.el7 rhel-7-server-extras-rpms 153 k\n librepo x86_64 1.8.1-8.el7_9 rhel 82 k\n libreport-filesystem\n x86_64 2.1.11-53.el7 rhel 41 k\n librhsm x86_64 0.0.3-3.el7_9 rhel-7-server-extras-rpms 28 k\n libsolv x86_64 0.6.34-4.el7 rhel 329 k\n lsof x86_64 4.87-6.el7 rhel 331 k\n lvm2 x86_64 7:2.02.187-6.el7_9.5 rhel 1.3 M\n lvm2-libs x86_64 7:2.02.187-6.el7_9.5 rhel 1.1 M\n mdadm x86_64 4.1-9.el7_9 rhel 440 k\n mpfr x86_64 3.1.1-4.el7 rhel 203 k\n pyparted x86_64 1:3.9-15.el7 rhel 195 k\n python2-blockdev x86_64 2.18-5.el7 rhel 61 k\n python2-bytesize x86_64 1.2-1.el7 rhel 22 k\n python2-hawkey x86_64 0.22.5-2.el7_9 rhel-7-server-extras-rpms 71 k\n python2-libdnf x86_64 0.22.5-2.el7_9 rhel-7-server-extras-rpms 611 k\n sgpio x86_64 1.2.0.10-13.el7 rhel 14 k\n volume_key-libs x86_64 0.3.9-9.el7 rhel 141 k\n\nTransaction Summary\n================================================================================\nInstall 7 Packages (+29 Dependent packages)\n\nTotal download size: 8.2 M\nInstalled size: 24 M\nDownloading packages:\n--------------------------------------------------------------------------------\nTotal 21 MB/s | 8.2 MB 00:00 \nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n Installing : libblockdev-utils-2.18-5.el7.x86_64 1/36 \n Installing : 7:device-mapper-event-libs-1.02.170-6.el7_9.5.x86_64 2/36 \n Installing : json-glib-1.4.2-2.el7.x86_64 3/36 \n Installing : librhsm-0.0.3-3.el7_9.x86_64 4/36 \n Installing : libsolv-0.6.34-4.el7.x86_64 5/36 \n Installing : libaio-0.3.109-13.el7.x86_64 6/36 \n Installing : librepo-1.8.1-8.el7_9.x86_64 7/36 \n Installing : libmodulemd-1.6.3-1.el7.x86_64 8/36 \n Installing : libdnf-0.22.5-2.el7_9.x86_64 9/36 \n Installing : device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64 10/36 \n Installing : 7:device-mapper-event-1.02.170-6.el7_9.5.x86_64 11/36 \n Installing : 7:lvm2-libs-2.02.187-6.el7_9.5.x86_64 12/36 \n Installing : 7:lvm2-2.02.187-6.el7_9.5.x86_64 13/36 \n Installing : python2-libdnf-0.22.5-2.el7_9.x86_64 14/36 \n Installing : python2-hawkey-0.22.5-2.el7_9.x86_64 15/36 \n Installing : libblockdev-2.18-5.el7.x86_64 16/36 \n Installing : python2-blockdev-2.18-5.el7.x86_64 17/36 \n Installing : 1:pyparted-3.9-15.el7.x86_64 18/36 \n Installing : sgpio-1.2.0.10-13.el7.x86_64 19/36 \n Installing : dmraid-1.0.0.rc16-28.el7.x86_64 20/36 \n Installing : dmraid-events-1.0.0.rc16-28.el7.x86_64 21/36 \n Installing : volume_key-libs-0.3.9-9.el7.x86_64 22/36 \n Installing : mpfr-3.1.1-4.el7.x86_64 23/36 \n Installing : libbytesize-1.2-1.el7.x86_64 24/36 \n Installing : python2-bytesize-1.2-1.el7.x86_64 25/36 \n Installing : libreport-filesystem-2.1.11-53.el7.x86_64 26/36 \n Installing : mdadm-4.1-9.el7_9.x86_64 27/36 \n Installing : 1:blivet3-data-3.1.3-3.el7.noarch 28/36 \n Installing : lsof-4.87-6.el7.x86_64 29/36 \n Installing : 1:python2-blivet3-3.1.3-3.el7.noarch 30/36 \n Installing : libblockdev-mdraid-2.18-5.el7.x86_64 31/36 \n Installing : libblockdev-crypto-2.18-5.el7.x86_64 32/36 \n Installing : libblockdev-dm-2.18-5.el7.x86_64 33/36 \n Installing : libblockdev-lvm-2.18-5.el7.x86_64 34/36 \n Installing : libblockdev-swap-2.18-5.el7.x86_64 35/36 \n Installing : python-enum34-1.0.4-1.el7.noarch 36/36 \n Verifying : 7:device-mapper-event-1.02.170-6.el7_9.5.x86_64 1/36 \n Verifying : libblockdev-swap-2.18-5.el7.x86_64 2/36 \n Verifying : librhsm-0.0.3-3.el7_9.x86_64 3/36 \n Verifying : libblockdev-lvm-2.18-5.el7.x86_64 4/36 \n Verifying : lsof-4.87-6.el7.x86_64 5/36 \n Verifying : libblockdev-mdraid-2.18-5.el7.x86_64 6/36 \n Verifying : libdnf-0.22.5-2.el7_9.x86_64 7/36 \n Verifying : python-enum34-1.0.4-1.el7.noarch 8/36 \n Verifying : 1:blivet3-data-3.1.3-3.el7.noarch 9/36 \n Verifying : dmraid-events-1.0.0.rc16-28.el7.x86_64 10/36 \n Verifying : python2-blockdev-2.18-5.el7.x86_64 11/36 \n Verifying : libmodulemd-1.6.3-1.el7.x86_64 12/36 \n Verifying : librepo-1.8.1-8.el7_9.x86_64 13/36 \n Verifying : libblockdev-dm-2.18-5.el7.x86_64 14/36 \n Verifying : json-glib-1.4.2-2.el7.x86_64 15/36 \n Verifying : libaio-0.3.109-13.el7.x86_64 16/36 \n Verifying : 7:lvm2-libs-2.02.187-6.el7_9.5.x86_64 17/36 \n Verifying : python2-hawkey-0.22.5-2.el7_9.x86_64 18/36 \n Verifying : python2-bytesize-1.2-1.el7.x86_64 19/36 \n Verifying : libblockdev-2.18-5.el7.x86_64 20/36 \n Verifying : libreport-filesystem-2.1.11-53.el7.x86_64 21/36 \n Verifying : libbytesize-1.2-1.el7.x86_64 22/36 \n Verifying : 7:device-mapper-event-libs-1.02.170-6.el7_9.5.x86_64 23/36 \n Verifying : python2-libdnf-0.22.5-2.el7_9.x86_64 24/36 \n Verifying : 7:lvm2-2.02.187-6.el7_9.5.x86_64 25/36 \n Verifying : libblockdev-utils-2.18-5.el7.x86_64 26/36 \n Verifying : mpfr-3.1.1-4.el7.x86_64 27/36 \n Verifying : volume_key-libs-0.3.9-9.el7.x86_64 28/36 \n Verifying : libsolv-0.6.34-4.el7.x86_64 29/36 \n Verifying : device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64 30/36 \n Verifying : 1:python2-blivet3-3.1.3-3.el7.noarch 31/36 \n Verifying : dmraid-1.0.0.rc16-28.el7.x86_64 32/36 \n Verifying : mdadm-4.1-9.el7_9.x86_64 33/36 \n Verifying : sgpio-1.2.0.10-13.el7.x86_64 34/36 \n Verifying : libblockdev-crypto-2.18-5.el7.x86_64 35/36 \n Verifying : 1:pyparted-3.9-15.el7.x86_64 36/36 \n\nInstalled:\n libblockdev-crypto.x86_64 0:2.18-5.el7 libblockdev-dm.x86_64 0:2.18-5.el7 \n libblockdev-lvm.x86_64 0:2.18-5.el7 libblockdev-mdraid.x86_64 0:2.18-5.el7\n libblockdev-swap.x86_64 0:2.18-5.el7 python-enum34.noarch 0:1.0.4-1.el7 \n python2-blivet3.noarch 1:3.1.3-3.el7 \n\nDependency Installed:\n blivet3-data.noarch 1:3.1.3-3.el7 \n device-mapper-event.x86_64 7:1.02.170-6.el7_9.5 \n device-mapper-event-libs.x86_64 7:1.02.170-6.el7_9.5 \n device-mapper-persistent-data.x86_64 0:0.8.5-3.el7_9.2 \n dmraid.x86_64 0:1.0.0.rc16-28.el7 \n dmraid-events.x86_64 0:1.0.0.rc16-28.el7 \n json-glib.x86_64 0:1.4.2-2.el7 \n libaio.x86_64 0:0.3.109-13.el7 \n libblockdev.x86_64 0:2.18-5.el7 \n libblockdev-utils.x86_64 0:2.18-5.el7 \n libbytesize.x86_64 0:1.2-1.el7 \n libdnf.x86_64 0:0.22.5-2.el7_9 \n libmodulemd.x86_64 0:1.6.3-1.el7 \n librepo.x86_64 0:1.8.1-8.el7_9 \n libreport-filesystem.x86_64 0:2.1.11-53.el7 \n librhsm.x86_64 0:0.0.3-3.el7_9 \n libsolv.x86_64 0:0.6.34-4.el7 \n lsof.x86_64 0:4.87-6.el7 \n lvm2.x86_64 7:2.02.187-6.el7_9.5 \n lvm2-libs.x86_64 7:2.02.187-6.el7_9.5 \n mdadm.x86_64 0:4.1-9.el7_9 \n mpfr.x86_64 0:3.1.1-4.el7 \n pyparted.x86_64 1:3.9-15.el7 \n python2-blockdev.x86_64 0:2.18-5.el7 \n python2-bytesize.x86_64 0:1.2-1.el7 \n python2-hawkey.x86_64 0:0.22.5-2.el7_9 \n python2-libdnf.x86_64 0:0.22.5-2.el7_9 \n sgpio.x86_64 0:1.2.0.10-13.el7 \n volume_key-libs.x86_64 0:0.3.9-9.el7 \n\nComplete!\n" ] } TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 15:24:20 +0000 (0:00:08.810) 0:00:12.734 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 15:24:20 +0000 (0:00:00.072) 0:00:12.806 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 15:24:20 +0000 (0:00:00.063) 0:00:12.869 ********* ok: [/cache/rhel-7.qcow2] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 15:24:21 +0000 (0:00:00.655) 0:00:13.525 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 15:24:21 +0000 (0:00:00.042) 0:00:13.568 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 15:24:21 +0000 (0:00:00.033) 0:00:13.602 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 15:24:21 +0000 (0:00:00.036) 0:00:13.639 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 15:24:21 +0000 (0:00:00.032) 0:00:13.671 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [] } TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 15:24:22 +0000 (0:00:00.590) 0:00:14.262 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "brandbot.service": { "name": "brandbot.service", "source": "systemd", "state": "stopped", "status": "static" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-shell.service": { "name": "console-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.import1.service": { "name": "dbus-org.freedesktop.import1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.machine1.service": { "name": "dbus-org.freedesktop.machine1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dmraid-activation.service": { "name": "dmraid-activation.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "exim.service": { "name": "exim.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "ip6tables.service": { "name": "ip6tables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iptables.service": { "name": "iptables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmetad.service": { "name": "lvm2-lvmetad.service", "source": "systemd", "state": "running", "status": "static" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "netconsole": { "name": "netconsole", "source": "sysv", "state": "stopped", "status": "disabled" }, "network": { "name": "network", "source": "sysv", "state": "running", "status": "enabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "active" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-config.service": { "name": "nfs-config.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-idmap.service": { "name": "nfs-idmap.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-lock.service": { "name": "nfs-lock.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-rquotad.service": { "name": "nfs-rquotad.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-secure-server.service": { "name": "nfs-secure-server.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-secure.service": { "name": "nfs-secure.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs.service": { "name": "nfs.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfslock.service": { "name": "nfslock.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ovirt-guest-agent.service": { "name": "ovirt-guest-agent.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "postfix.service": { "name": "postfix.service", "source": "systemd", "state": "running", "status": "enabled" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhel-autorelabel-mark.service": { "name": "rhel-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-autorelabel.service": { "name": "rhel-autorelabel.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-configure.service": { "name": "rhel-configure.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-dmesg.service": { "name": "rhel-dmesg.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-domainname.service": { "name": "rhel-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-import-state.service": { "name": "rhel-import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-loadmodules.service": { "name": "rhel-loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-readonly.service": { "name": "rhel-readonly.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhnsd": { "name": "rhnsd", "source": "sysv", "state": "running", "status": "enabled" }, "rhnsd.service": { "name": "rhnsd.service", "source": "systemd", "state": "running", "status": "active" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-rquotad.service": { "name": "rpc-rquotad.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpcgssd.service": { "name": "rpcgssd.service", "source": "systemd", "state": "inactive", "status": "static" }, "rpcidmapd.service": { "name": "rpcidmapd.service", "source": "systemd", "state": "inactive", "status": "static" }, "rsyncd.service": { "name": "rsyncd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyncd@.service": { "name": "rsyncd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-policy-migrate-local-changes@.service": { "name": "selinux-policy-migrate-local-changes@.service", "source": "systemd", "state": "unknown", "status": "static" }, "selinux-policy-migrate-local-changes@targeted.service": { "name": "selinux-policy-migrate-local-changes@targeted.service", "source": "systemd", "state": "stopped", "status": "active" }, "sendmail.service": { "name": "sendmail.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "static" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bootchart.service": { "name": "systemd-bootchart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-importd.service": { "name": "systemd-importd.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machined.service": { "name": "systemd-machined.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-nspawn@.service": { "name": "systemd-nspawn@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-readahead-collect.service": { "name": "systemd-readahead-collect.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-readahead-done.service": { "name": "systemd-readahead-done.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "systemd-readahead-drop.service": { "name": "systemd-readahead-drop.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "systemd-readahead-replay.service": { "name": "systemd-readahead-replay.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill@.service": { "name": "systemd-rfkill@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-shutdownd.service": { "name": "systemd-shutdownd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "wpa_supplicant.service": { "name": "wpa_supplicant.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 15:24:23 +0000 (0:00:01.174) 0:00:15.436 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 15:24:23 +0000 (0:00:00.052) 0:00:15.489 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 15:24:23 +0000 (0:00:00.020) 0:00:15.509 ********* ok: [/cache/rhel-7.qcow2] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 15:24:23 +0000 (0:00:00.471) 0:00:15.981 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 15:24:24 +0000 (0:00:00.038) 0:00:16.019 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 15:24:24 +0000 (0:00:00.022) 0:00:16.041 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 15:24:24 +0000 (0:00:00.038) 0:00:16.080 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 15:24:24 +0000 (0:00:00.033) 0:00:16.113 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 15:24:24 +0000 (0:00:00.032) 0:00:16.146 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 15:24:24 +0000 (0:00:00.032) 0:00:16.178 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 15:24:24 +0000 (0:00:00.022) 0:00:16.201 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 15:24:24 +0000 (0:00:00.032) 0:00:16.233 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 15:24:24 +0000 (0:00:00.022) 0:00:16.256 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658417058.2394526, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658201031.524, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 70, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658200515.884, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744071677828413", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 15:24:24 +0000 (0:00:00.426) 0:00:16.683 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 15:24:24 +0000 (0:00:00.021) 0:00:16.704 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [Mark tasks to be skipped] ************************************************ task path: /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes.yml:14 Thursday 21 July 2022 15:24:25 +0000 (0:00:00.869) 0:00:17.574 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_skip_checks": [ "blivet_available", "packages_installed", "service_facts" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes.yml:21 Thursday 21 July 2022 15:24:25 +0000 (0:00:00.060) 0:00:17.635 ********* included: /tmp/tmpaxjje44y/tests/get_unused_disk.yml for /cache/rhel-7.qcow2 TASK [Find unused disks in the system] ***************************************** task path: /tmp/tmpaxjje44y/tests/get_unused_disk.yml:2 Thursday 21 July 2022 15:24:25 +0000 (0:00:00.060) 0:00:17.695 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "disks": [ "sdb" ] } TASK [Set unused_disks if necessary] ******************************************* task path: /tmp/tmpaxjje44y/tests/get_unused_disk.yml:9 Thursday 21 July 2022 15:24:26 +0000 (0:00:00.459) 0:00:18.154 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "unused_disks": [ "sdb" ] }, "changed": false } TASK [Exit playbook when there's not enough unused disks in the system] ******** task path: /tmp/tmpaxjje44y/tests/get_unused_disk.yml:14 Thursday 21 July 2022 15:24:26 +0000 (0:00:00.037) 0:00:18.192 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Print unused disks] ****************************************************** task path: /tmp/tmpaxjje44y/tests/get_unused_disk.yml:19 Thursday 21 July 2022 15:24:26 +0000 (0:00:00.040) 0:00:18.232 ********* ok: [/cache/rhel-7.qcow2] => { "unused_disks": [ "sdb" ] } TASK [Create three LVM logical volumes under one volume group] ***************** task path: /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes.yml:26 Thursday 21 July 2022 15:24:26 +0000 (0:00:00.036) 0:00:18.268 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 15:24:26 +0000 (0:00:00.044) 0:00:18.312 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 15:24:26 +0000 (0:00:00.033) 0:00:18.346 ********* ok: [/cache/rhel-7.qcow2] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 15:24:26 +0000 (0:00:00.406) 0:00:18.752 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 15:24:26 +0000 (0:00:00.076) 0:00:18.828 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 15:24:26 +0000 (0:00:00.036) 0:00:18.865 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 15:24:26 +0000 (0:00:00.035) 0:00:18.900 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 15:24:26 +0000 (0:00:00.044) 0:00:18.945 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 15:24:26 +0000 (0:00:00.019) 0:00:18.965 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 15:24:27 +0000 (0:00:00.025) 0:00:18.990 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "sdb" ], "name": "foo", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "3g" }, { "mount_point": "/opt/test2", "name": "test2", "size": "3g" }, { "mount_point": "/opt/test3", "name": "test3", "size": "3g" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 15:24:27 +0000 (0:00:00.034) 0:00:19.025 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 15:24:27 +0000 (0:00:00.037) 0:00:19.062 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 15:24:27 +0000 (0:00:00.027) 0:00:19.090 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 15:24:27 +0000 (0:00:00.028) 0:00:19.119 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 15:24:27 +0000 (0:00:00.027) 0:00:19.147 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 15:24:27 +0000 (0:00:00.026) 0:00:19.173 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 15:24:27 +0000 (0:00:00.053) 0:00:19.227 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 15:24:27 +0000 (0:00:00.021) 0:00:19.248 ********* changed: [/cache/rhel-7.qcow2] => { "actions": [ { "action": "create format", "device": "/dev/sdb", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test3", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test3", "fs_type": "xfs" }, { "action": "create device", "device": "/dev/mapper/foo-test2", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test2", "fs_type": "xfs" }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2", "/dev/mapper/foo-test3" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 15:24:29 +0000 (0:00:01.892) 0:00:21.141 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 15:24:29 +0000 (0:00:00.035) 0:00:21.176 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 15:24:29 +0000 (0:00:00.048) 0:00:21.225 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/sdb", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test3", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test3", "fs_type": "xfs" }, { "action": "create device", "device": "/dev/mapper/foo-test2", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test2", "fs_type": "xfs" }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2", "/dev/mapper/foo-test3" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 15:24:29 +0000 (0:00:00.039) 0:00:21.264 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 15:24:29 +0000 (0:00:00.040) 0:00:21.305 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 15:24:29 +0000 (0:00:00.037) 0:00:21.342 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 15:24:29 +0000 (0:00:00.040) 0:00:21.382 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 15:24:30 +0000 (0:00:00.710) 0:00:22.093 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test2', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test2', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test3', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test3', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "mounted" }, "name": "/opt/test3", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test3" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 15:24:31 +0000 (0:00:01.113) 0:00:23.207 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 15:24:31 +0000 (0:00:00.471) 0:00:23.678 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658417058.2394526, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658201031.524, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 70, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658200515.884, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744071677828413", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 15:24:32 +0000 (0:00:00.325) 0:00:24.004 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 15:24:32 +0000 (0:00:00.023) 0:00:24.027 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes.yml:44 Thursday 21 July 2022 15:24:32 +0000 (0:00:00.849) 0:00:24.877 ********* included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1 Thursday 21 July 2022 15:24:32 +0000 (0:00:00.040) 0:00:24.918 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6 Thursday 21 July 2022 15:24:33 +0000 (0:00:00.084) 0:00:25.002 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14 Thursday 21 July 2022 15:24:33 +0000 (0:00:00.037) 0:00:25.040 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "3G", "type": "lvm", "uuid": "a47eb0d6-9f88-47db-870b-ef3b58e60afc" }, "/dev/mapper/foo-test2": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test2", "size": "3G", "type": "lvm", "uuid": "2f03fe81-8a5e-4cf6-a9d1-f783f730dd35" }, "/dev/mapper/foo-test3": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test3", "size": "3G", "type": "lvm", "uuid": "3f441d29-fa52-465c-934e-baef0ec1af88" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "c1Mk0I-V3FP-Vj0J-xjuH-291X-Hwf5-bK1aAM" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-23-55-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19 Thursday 21 July 2022 15:24:33 +0000 (0:00:00.439) 0:00:25.479 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003946", "end": "2022-07-21 11:24:33.624151", "rc": 0, "start": "2022-07-21 11:24:33.620205" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 /dev/mapper/foo-test2 /opt/test2 xfs defaults 0 0 /dev/mapper/foo-test3 /opt/test3 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24 Thursday 21 July 2022 15:24:33 +0000 (0:00:00.432) 0:00:25.911 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003460", "end": "2022-07-21 11:24:33.958060", "failed_when_result": false, "rc": 0, "start": "2022-07-21 11:24:33.954600" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33 Thursday 21 July 2022 15:24:34 +0000 (0:00:00.332) 0:00:26.244 ********* included: /tmp/tmpaxjje44y/tests/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, {'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, {'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:5 Thursday 21 July 2022 15:24:34 +0000 (0:00:00.062) 0:00:26.307 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:18 Thursday 21 July 2022 15:24:34 +0000 (0:00:00.035) 0:00:26.343 ********* included: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members) included: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 15:24:34 +0000 (0:00:00.047) 0:00:26.390 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sdb" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 15:24:34 +0000 (0:00:00.055) 0:00:26.445 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb", "pv": "/dev/sdb" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 15:24:34 +0000 (0:00:00.418) 0:00:26.863 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 15:24:34 +0000 (0:00:00.059) 0:00:26.923 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sdb" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 15:24:34 +0000 (0:00:00.056) 0:00:26.980 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.055) 0:00:27.035 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.041) 0:00:27.077 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.050) 0:00:27.127 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.026) 0:00:27.153 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.046) 0:00:27.200 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-md.yml for /cache/rhel-7.qcow2 TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:6 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.043) 0:00:27.243 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:12 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.026) 0:00:27.269 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:16 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.026) 0:00:27.296 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:20 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.024) 0:00:27.320 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:24 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.025) 0:00:27.346 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:30 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.024) 0:00:27.371 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:36 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.025) 0:00:27.396 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:44 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.024) 0:00:27.420 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.034) 0:00:27.454 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.043) 0:00:27.498 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.054) 0:00:27.553 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.031) 0:00:27.585 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.027) 0:00:27.613 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.029) 0:00:27.642 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.027) 0:00:27.670 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.070) 0:00:27.741 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.028) 0:00:27.769 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.027) 0:00:27.796 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.027) 0:00:27.824 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.027) 0:00:27.852 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml for /cache/rhel-7.qcow2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.043) 0:00:27.896 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.065) 0:00:27.961 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:24:35 +0000 (0:00:00.024) 0:00:27.985 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.024) 0:00:28.009 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.025) 0:00:28.034 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.023) 0:00:28.057 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.023) 0:00:28.081 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.023) 0:00:28.104 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.024) 0:00:28.128 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.023) 0:00:28.152 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.023) 0:00:28.175 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.022) 0:00:28.198 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.023) 0:00:28.222 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.022) 0:00:28.245 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2 TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.046) 0:00:28.291 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.053) 0:00:28.344 ********* skipping: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "_storage_test_pool_member_path": "/dev/sdb", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.028) 0:00:28.373 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/sdb) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.044) 0:00:28.417 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.049) 0:00:28.467 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.049) 0:00:28.517 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.036) 0:00:28.553 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.039) 0:00:28.593 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.036) 0:00:28.629 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.033) 0:00:28.663 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.035) 0:00:28.699 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.048) 0:00:28.748 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.065) 0:00:28.813 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.030) 0:00:28.844 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.025) 0:00:28.869 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.026) 0:00:28.896 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.053) 0:00:28.950 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:24:36 +0000 (0:00:00.024) 0:00:28.974 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.026) 0:00:29.000 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.025) 0:00:29.026 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.036) 0:00:29.062 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.025) 0:00:29.087 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.025) 0:00:29.113 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.024) 0:00:29.138 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.026) 0:00:29.165 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.024) 0:00:29.189 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.024) 0:00:29.213 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.023) 0:00:29.237 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.036) 0:00:29.273 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.025) 0:00:29.299 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.026) 0:00:29.326 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.024) 0:00:29.351 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.026) 0:00:29.377 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.023) 0:00:29.400 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.023) 0:00:29.424 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.024) 0:00:29.448 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.036) 0:00:29.484 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.033) 0:00:29.518 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.051) 0:00:29.569 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.050) 0:00:29.619 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.085) 0:00:29.705 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.042) 0:00:29.748 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "a47eb0d6-9f88-47db-870b-ef3b58e60afc" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "a47eb0d6-9f88-47db-870b-ef3b58e60afc" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.055) 0:00:29.804 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.052) 0:00:29.856 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.052) 0:00:29.908 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.051) 0:00:29.959 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:24:37 +0000 (0:00:00.024) 0:00:29.983 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.023) 0:00:30.007 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.023) 0:00:30.030 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.033) 0:00:30.064 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.097) 0:00:30.162 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.139) 0:00:30.302 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.052) 0:00:30.354 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.047) 0:00:30.402 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.044) 0:00:30.446 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.040) 0:00:30.487 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.042) 0:00:30.529 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658417068.8004527, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658417068.8004527, "dev": 5, "device_type": 64514, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 32400, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658417068.8004527, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.320) 0:00:30.850 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.037) 0:00:30.888 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.038) 0:00:30.927 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.036) 0:00:30.963 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:24:38 +0000 (0:00:00.023) 0:00:30.987 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:24:39 +0000 (0:00:00.042) 0:00:31.029 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:24:39 +0000 (0:00:00.027) 0:00:31.056 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "changes": { "installed": [ "cryptsetup" ] }, "rc": 0, "results": [ "Loaded plugins: search-disabled-repos\nResolving Dependencies\n--> Running transaction check\n---> Package cryptsetup.x86_64 0:2.0.3-6.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package Arch Version Repository Size\n================================================================================\nInstalling:\n cryptsetup x86_64 2.0.3-6.el7 rhel 154 k\n\nTransaction Summary\n================================================================================\nInstall 1 Package\n\nTotal download size: 154 k\nInstalled size: 354 k\nDownloading packages:\nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n Installing : cryptsetup-2.0.3-6.el7.x86_64 1/1 \n Verifying : cryptsetup-2.0.3-6.el7.x86_64 1/1 \n\nInstalled:\n cryptsetup.x86_64 0:2.0.3-6.el7 \n\nComplete!\n" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:24:40 +0000 (0:00:01.529) 0:00:32.586 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:24:40 +0000 (0:00:00.024) 0:00:32.611 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:24:40 +0000 (0:00:00.023) 0:00:32.634 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:24:40 +0000 (0:00:00.052) 0:00:32.687 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:24:40 +0000 (0:00:00.023) 0:00:32.710 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:24:40 +0000 (0:00:00.022) 0:00:32.732 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:24:40 +0000 (0:00:00.022) 0:00:32.755 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:24:40 +0000 (0:00:00.023) 0:00:32.779 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:24:40 +0000 (0:00:00.022) 0:00:32.802 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:24:40 +0000 (0:00:00.054) 0:00:32.857 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:24:40 +0000 (0:00:00.052) 0:00:32.909 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:24:40 +0000 (0:00:00.038) 0:00:32.947 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:24:40 +0000 (0:00:00.036) 0:00:32.984 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.035) 0:00:33.020 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.038) 0:00:33.058 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.037) 0:00:33.096 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.036) 0:00:33.132 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.037) 0:00:33.170 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.047) 0:00:33.218 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.040) 0:00:33.258 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.038) 0:00:33.297 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.039) 0:00:33.336 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:24:41 +0000 (0:00:00.439) 0:00:33.775 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.358) 0:00:34.134 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.125) 0:00:34.260 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.035) 0:00:34.296 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.039) 0:00:34.336 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.041) 0:00:34.378 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.041) 0:00:34.419 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.037) 0:00:34.456 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.038) 0:00:34.495 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.038) 0:00:34.534 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.041) 0:00:34.575 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.053) 0:00:34.629 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.023252", "end": "2022-07-21 11:24:42.679556", "rc": 0, "start": "2022-07-21 11:24:42.656304" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:24:42 +0000 (0:00:00.339) 0:00:34.969 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.049) 0:00:35.019 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.048) 0:00:35.068 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.036) 0:00:35.104 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.038) 0:00:35.143 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.048) 0:00:35.192 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.042) 0:00:35.235 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.038) 0:00:35.273 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.052) 0:00:35.326 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.085) 0:00:35.411 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.043) 0:00:35.455 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test2", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "2f03fe81-8a5e-4cf6-a9d1-f783f730dd35" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test2", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "2f03fe81-8a5e-4cf6-a9d1-f783f730dd35" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.058) 0:00:35.513 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.052) 0:00:35.566 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.053) 0:00:35.619 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.050) 0:00:35.669 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.026) 0:00:35.695 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.027) 0:00:35.723 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.026) 0:00:35.749 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.037) 0:00:35.786 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test2 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.061) 0:00:35.848 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.049) 0:00:35.897 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:24:43 +0000 (0:00:00.081) 0:00:35.979 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:24:44 +0000 (0:00:00.035) 0:00:36.014 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:24:44 +0000 (0:00:00.102) 0:00:36.116 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:24:44 +0000 (0:00:00.039) 0:00:36.156 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:24:44 +0000 (0:00:00.041) 0:00:36.197 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658417068.6214526, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658417068.6214526, "dev": 5, "device_type": 64513, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 32354, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658417068.6214526, "nlink": 1, "path": "/dev/mapper/foo-test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:24:44 +0000 (0:00:00.319) 0:00:36.516 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:24:44 +0000 (0:00:00.037) 0:00:36.554 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:24:44 +0000 (0:00:00.038) 0:00:36.592 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:24:44 +0000 (0:00:00.039) 0:00:36.631 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:24:44 +0000 (0:00:00.025) 0:00:36.657 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:24:44 +0000 (0:00:00.040) 0:00:36.697 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:24:44 +0000 (0:00:00.024) 0:00:36.722 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.521) 0:00:37.244 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.026) 0:00:37.271 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.025) 0:00:37.296 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.053) 0:00:37.350 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.025) 0:00:37.375 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.023) 0:00:37.399 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.023) 0:00:37.423 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.024) 0:00:37.448 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.023) 0:00:37.471 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.052) 0:00:37.524 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.049) 0:00:37.573 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.036) 0:00:37.609 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.037) 0:00:37.647 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.035) 0:00:37.682 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.036) 0:00:37.719 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.035) 0:00:37.754 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.037) 0:00:37.792 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.036) 0:00:37.829 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.038) 0:00:37.867 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.035) 0:00:37.902 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.036) 0:00:37.939 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:24:45 +0000 (0:00:00.035) 0:00:37.974 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:24:46 +0000 (0:00:00.348) 0:00:38.322 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:24:46 +0000 (0:00:00.358) 0:00:38.680 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:24:46 +0000 (0:00:00.052) 0:00:38.733 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:24:46 +0000 (0:00:00.036) 0:00:38.770 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:24:46 +0000 (0:00:00.040) 0:00:38.810 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:24:46 +0000 (0:00:00.039) 0:00:38.849 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:24:46 +0000 (0:00:00.036) 0:00:38.886 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:24:46 +0000 (0:00:00.035) 0:00:38.922 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:24:46 +0000 (0:00:00.034) 0:00:38.957 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.036) 0:00:38.993 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.034) 0:00:39.028 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.055) 0:00:39.083 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test2" ], "delta": "0:00:00.023950", "end": "2022-07-21 11:24:47.128093", "rc": 0, "start": "2022-07-21 11:24:47.104143" } STDOUT: LVM2_LV_NAME=test2 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.335) 0:00:39.419 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.052) 0:00:39.472 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.052) 0:00:39.524 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.039) 0:00:39.563 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.040) 0:00:39.604 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.038) 0:00:39.643 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.038) 0:00:39.681 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.038) 0:00:39.719 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.052) 0:00:39.772 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.076) 0:00:39.849 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test3" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.039) 0:00:39.888 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test3", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test3", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "3f441d29-fa52-465c-934e-baef0ec1af88" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test3", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test3", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "3f441d29-fa52-465c-934e-baef0ec1af88" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:24:47 +0000 (0:00:00.053) 0:00:39.942 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.049) 0:00:39.991 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.048) 0:00:40.039 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.052) 0:00:40.092 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.026) 0:00:40.118 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.024) 0:00:40.143 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.025) 0:00:40.168 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.069) 0:00:40.237 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test3 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test3 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test3 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.063) 0:00:40.301 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.090) 0:00:40.392 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.081) 0:00:40.473 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.038) 0:00:40.512 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.070) 0:00:40.583 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.115) 0:00:40.698 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:24:48 +0000 (0:00:00.038) 0:00:40.737 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658417068.4414525, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658417068.4414525, "dev": 5, "device_type": 64512, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 32309, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658417068.4414525, "nlink": 1, "path": "/dev/mapper/foo-test3", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.318) 0:00:41.056 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.043) 0:00:41.100 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.041) 0:00:41.141 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.037) 0:00:41.179 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.026) 0:00:41.205 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.039) 0:00:41.244 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.027) 0:00:41.272 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.521) 0:00:41.793 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.025) 0:00:41.819 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.025) 0:00:41.845 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.052) 0:00:41.897 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.028) 0:00:41.926 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.030) 0:00:41.956 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:24:49 +0000 (0:00:00.026) 0:00:41.983 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.028) 0:00:42.011 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.024) 0:00:42.036 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.054) 0:00:42.090 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.050) 0:00:42.141 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.037) 0:00:42.179 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.043) 0:00:42.222 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.038) 0:00:42.261 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.034) 0:00:42.295 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.034) 0:00:42.330 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.041) 0:00:42.371 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.037) 0:00:42.409 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.039) 0:00:42.448 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.039) 0:00:42.488 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.039) 0:00:42.527 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.038) 0:00:42.566 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:24:50 +0000 (0:00:00.366) 0:00:42.932 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:24:51 +0000 (0:00:00.367) 0:00:43.300 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:24:51 +0000 (0:00:00.084) 0:00:43.385 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:24:51 +0000 (0:00:00.037) 0:00:43.422 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:24:51 +0000 (0:00:00.039) 0:00:43.461 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:24:51 +0000 (0:00:00.040) 0:00:43.502 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:24:51 +0000 (0:00:00.038) 0:00:43.540 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:24:51 +0000 (0:00:00.038) 0:00:43.579 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:24:51 +0000 (0:00:00.038) 0:00:43.617 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:24:51 +0000 (0:00:00.038) 0:00:43.656 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:24:51 +0000 (0:00:00.033) 0:00:43.690 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:24:51 +0000 (0:00:00.053) 0:00:43.743 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test3" ], "delta": "0:00:00.028089", "end": "2022-07-21 11:24:51.796760", "rc": 0, "start": "2022-07-21 11:24:51.768671" } STDOUT: LVM2_LV_NAME=test3 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.343) 0:00:44.086 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.051) 0:00:44.138 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.051) 0:00:44.190 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.039) 0:00:44.229 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.039) 0:00:44.269 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.037) 0:00:44.307 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.039) 0:00:44.346 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.036) 0:00:44.383 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.023) 0:00:44.407 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Repeat the previous invocation to verify idempotence] ******************** task path: /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes.yml:46 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.035) 0:00:44.442 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.047) 0:00:44.490 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.038) 0:00:44.528 ********* ok: [/cache/rhel-7.qcow2] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 15:24:52 +0000 (0:00:00.419) 0:00:44.948 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 15:24:53 +0000 (0:00:00.062) 0:00:45.010 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 15:24:53 +0000 (0:00:00.034) 0:00:45.044 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 15:24:53 +0000 (0:00:00.033) 0:00:45.078 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 15:24:53 +0000 (0:00:00.046) 0:00:45.125 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 15:24:53 +0000 (0:00:00.022) 0:00:45.147 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 15:24:53 +0000 (0:00:00.030) 0:00:45.177 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "sdb" ], "name": "foo", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "3g" }, { "mount_point": "/opt/test2", "name": "test2", "size": "3g" }, { "mount_point": "/opt/test3", "name": "test3", "size": "3g" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 15:24:53 +0000 (0:00:00.094) 0:00:45.272 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 15:24:53 +0000 (0:00:00.038) 0:00:45.310 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 15:24:53 +0000 (0:00:00.028) 0:00:45.339 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 15:24:53 +0000 (0:00:00.030) 0:00:45.369 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 15:24:53 +0000 (0:00:00.030) 0:00:45.400 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 15:24:53 +0000 (0:00:00.028) 0:00:45.428 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 15:24:53 +0000 (0:00:00.059) 0:00:45.488 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 15:24:53 +0000 (0:00:00.023) 0:00:45.512 ********* ok: [/cache/rhel-7.qcow2] => { "actions": [], "changed": false, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/mapper/foo-test2", "/dev/mapper/foo-test3", "/dev/mapper/foo-test1", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 15:24:55 +0000 (0:00:01.721) 0:00:47.233 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 15:24:55 +0000 (0:00:00.038) 0:00:47.272 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 15:24:55 +0000 (0:00:00.023) 0:00:47.295 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/mapper/foo-test2", "/dev/mapper/foo-test3", "/dev/mapper/foo-test1", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 15:24:55 +0000 (0:00:00.042) 0:00:47.338 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 15:24:55 +0000 (0:00:00.047) 0:00:47.386 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 15:24:55 +0000 (0:00:00.046) 0:00:47.433 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 15:24:55 +0000 (0:00:00.038) 0:00:47.472 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 15:24:55 +0000 (0:00:00.461) 0:00:47.933 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test2', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test2', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test3', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test3', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "mounted" }, "name": "/opt/test3", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test3" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 15:24:56 +0000 (0:00:00.954) 0:00:48.888 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 15:24:57 +0000 (0:00:00.486) 0:00:49.374 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658417058.2394526, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658201031.524, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 70, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658200515.884, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744071677828413", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 15:24:57 +0000 (0:00:00.319) 0:00:49.694 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 15:24:57 +0000 (0:00:00.025) 0:00:49.720 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes.yml:64 Thursday 21 July 2022 15:24:58 +0000 (0:00:00.934) 0:00:50.654 ********* included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1 Thursday 21 July 2022 15:24:58 +0000 (0:00:00.044) 0:00:50.698 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6 Thursday 21 July 2022 15:24:58 +0000 (0:00:00.052) 0:00:50.751 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14 Thursday 21 July 2022 15:24:58 +0000 (0:00:00.039) 0:00:50.790 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "3G", "type": "lvm", "uuid": "a47eb0d6-9f88-47db-870b-ef3b58e60afc" }, "/dev/mapper/foo-test2": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test2", "size": "3G", "type": "lvm", "uuid": "2f03fe81-8a5e-4cf6-a9d1-f783f730dd35" }, "/dev/mapper/foo-test3": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test3", "size": "3G", "type": "lvm", "uuid": "3f441d29-fa52-465c-934e-baef0ec1af88" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "c1Mk0I-V3FP-Vj0J-xjuH-291X-Hwf5-bK1aAM" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-23-55-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19 Thursday 21 July 2022 15:24:59 +0000 (0:00:00.317) 0:00:51.108 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003482", "end": "2022-07-21 11:24:59.123636", "rc": 0, "start": "2022-07-21 11:24:59.120154" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 /dev/mapper/foo-test2 /opt/test2 xfs defaults 0 0 /dev/mapper/foo-test3 /opt/test3 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24 Thursday 21 July 2022 15:24:59 +0000 (0:00:00.304) 0:00:51.413 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003449", "end": "2022-07-21 11:24:59.435523", "failed_when_result": false, "rc": 0, "start": "2022-07-21 11:24:59.432074" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33 Thursday 21 July 2022 15:24:59 +0000 (0:00:00.312) 0:00:51.725 ********* included: /tmp/tmpaxjje44y/tests/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, {'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, {'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:5 Thursday 21 July 2022 15:24:59 +0000 (0:00:00.063) 0:00:51.789 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:18 Thursday 21 July 2022 15:24:59 +0000 (0:00:00.036) 0:00:51.826 ********* included: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members) included: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 15:24:59 +0000 (0:00:00.048) 0:00:51.874 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sdb" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 15:24:59 +0000 (0:00:00.054) 0:00:51.929 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb", "pv": "/dev/sdb" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.298) 0:00:52.227 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.048) 0:00:52.275 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sdb" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.052) 0:00:52.328 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.051) 0:00:52.379 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.039) 0:00:52.419 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.051) 0:00:52.470 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.024) 0:00:52.495 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.049) 0:00:52.544 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-md.yml for /cache/rhel-7.qcow2 TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:6 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.048) 0:00:52.593 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:12 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.024) 0:00:52.617 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:16 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.022) 0:00:52.640 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:20 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.024) 0:00:52.664 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:24 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.061) 0:00:52.726 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:30 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.025) 0:00:52.752 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:36 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.026) 0:00:52.779 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:44 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.026) 0:00:52.805 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.036) 0:00:52.842 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.042) 0:00:52.884 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.054) 0:00:52.938 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:25:00 +0000 (0:00:00.030) 0:00:52.969 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.027) 0:00:52.996 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.033) 0:00:53.029 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.029) 0:00:53.059 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.028) 0:00:53.087 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.027) 0:00:53.114 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.029) 0:00:53.143 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.025) 0:00:53.169 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.025) 0:00:53.195 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml for /cache/rhel-7.qcow2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.043) 0:00:53.238 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.057) 0:00:53.295 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.024) 0:00:53.320 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.025) 0:00:53.345 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.023) 0:00:53.369 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.022) 0:00:53.392 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.025) 0:00:53.417 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.025) 0:00:53.443 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.024) 0:00:53.467 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.025) 0:00:53.493 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.024) 0:00:53.517 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.024) 0:00:53.542 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.026) 0:00:53.568 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.025) 0:00:53.594 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2 TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.049) 0:00:53.644 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.051) 0:00:53.695 ********* skipping: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "_storage_test_pool_member_path": "/dev/sdb", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.031) 0:00:53.727 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/sdb) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.044) 0:00:53.771 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.090) 0:00:53.862 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.083) 0:00:53.945 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 15:25:01 +0000 (0:00:00.037) 0:00:53.983 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.038) 0:00:54.021 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.050) 0:00:54.071 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.037) 0:00:54.109 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.036) 0:00:54.145 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.050) 0:00:54.195 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.061) 0:00:54.257 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.026) 0:00:54.283 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.024) 0:00:54.308 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.024) 0:00:54.333 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.026) 0:00:54.360 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.027) 0:00:54.387 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.025) 0:00:54.413 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.024) 0:00:54.437 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.035) 0:00:54.473 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.028) 0:00:54.501 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.024) 0:00:54.526 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.023) 0:00:54.549 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.025) 0:00:54.575 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.028) 0:00:54.603 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.024) 0:00:54.628 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.022) 0:00:54.651 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.033) 0:00:54.684 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.026) 0:00:54.711 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.023) 0:00:54.735 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.022) 0:00:54.757 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.022) 0:00:54.780 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.023) 0:00:54.804 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.022) 0:00:54.826 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.022) 0:00:54.849 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.035) 0:00:54.884 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.037) 0:00:54.922 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:25:02 +0000 (0:00:00.051) 0:00:54.974 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.053) 0:00:55.027 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.117) 0:00:55.145 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.040) 0:00:55.186 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "a47eb0d6-9f88-47db-870b-ef3b58e60afc" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "a47eb0d6-9f88-47db-870b-ef3b58e60afc" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.057) 0:00:55.244 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.052) 0:00:55.297 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.067) 0:00:55.364 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.050) 0:00:55.414 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.023) 0:00:55.437 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.024) 0:00:55.462 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.022) 0:00:55.485 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.033) 0:00:55.519 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.059) 0:00:55.578 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.049) 0:00:55.628 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.052) 0:00:55.680 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.037) 0:00:55.718 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.039) 0:00:55.758 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.038) 0:00:55.796 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:25:03 +0000 (0:00:00.039) 0:00:55.836 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658417068.8004527, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658417068.8004527, "dev": 5, "device_type": 64514, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 32400, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658417068.8004527, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.351) 0:00:56.187 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.042) 0:00:56.230 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.040) 0:00:56.271 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.037) 0:00:56.308 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.023) 0:00:56.332 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.039) 0:00:56.371 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.026) 0:00:56.398 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.538) 0:00:56.937 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.024) 0:00:56.962 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:25:04 +0000 (0:00:00.026) 0:00:56.988 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.056) 0:00:57.044 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.026) 0:00:57.071 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.026) 0:00:57.098 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.026) 0:00:57.124 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.025) 0:00:57.150 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.022) 0:00:57.173 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.101) 0:00:57.274 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.054) 0:00:57.329 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.035) 0:00:57.364 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.034) 0:00:57.399 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.035) 0:00:57.434 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.077) 0:00:57.512 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.036) 0:00:57.549 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.035) 0:00:57.585 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.068) 0:00:57.654 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.036) 0:00:57.691 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.035) 0:00:57.726 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.035) 0:00:57.762 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:25:05 +0000 (0:00:00.045) 0:00:57.807 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.310) 0:00:58.118 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.325) 0:00:58.443 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.053) 0:00:58.496 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.035) 0:00:58.532 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.040) 0:00:58.572 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.039) 0:00:58.612 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.037) 0:00:58.649 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.038) 0:00:58.688 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.040) 0:00:58.728 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.038) 0:00:58.767 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.038) 0:00:58.805 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:25:06 +0000 (0:00:00.052) 0:00:58.858 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.028793", "end": "2022-07-21 11:25:06.910059", "rc": 0, "start": "2022-07-21 11:25:06.881266" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:25:07 +0000 (0:00:00.341) 0:00:59.199 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:25:07 +0000 (0:00:00.049) 0:00:59.249 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:25:07 +0000 (0:00:00.049) 0:00:59.299 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:25:07 +0000 (0:00:00.037) 0:00:59.337 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:25:07 +0000 (0:00:00.035) 0:00:59.372 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:25:07 +0000 (0:00:00.035) 0:00:59.408 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:25:07 +0000 (0:00:00.036) 0:00:59.444 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:25:07 +0000 (0:00:00.038) 0:00:59.483 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:25:07 +0000 (0:00:00.047) 0:00:59.531 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:25:07 +0000 (0:00:00.080) 0:00:59.611 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:25:07 +0000 (0:00:00.098) 0:00:59.710 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test2", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "2f03fe81-8a5e-4cf6-a9d1-f783f730dd35" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test2", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "2f03fe81-8a5e-4cf6-a9d1-f783f730dd35" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:25:07 +0000 (0:00:00.057) 0:00:59.767 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:25:07 +0000 (0:00:00.089) 0:00:59.857 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:25:07 +0000 (0:00:00.098) 0:00:59.956 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.100) 0:01:00.056 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.032) 0:01:00.088 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.068) 0:01:00.157 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.024) 0:01:00.182 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.034) 0:01:00.217 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test2 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.062) 0:01:00.279 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.054) 0:01:00.334 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.049) 0:01:00.383 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.039) 0:01:00.423 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.039) 0:01:00.462 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.049) 0:01:00.512 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.042) 0:01:00.555 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658417068.6214526, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658417068.6214526, "dev": 5, "device_type": 64513, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 32354, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658417068.6214526, "nlink": 1, "path": "/dev/mapper/foo-test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.321) 0:01:00.876 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.041) 0:01:00.918 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:25:08 +0000 (0:00:00.042) 0:01:00.960 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.041) 0:01:01.001 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.026) 0:01:01.027 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.044) 0:01:01.072 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.023) 0:01:01.096 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.531) 0:01:01.627 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.026) 0:01:01.654 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.025) 0:01:01.680 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.056) 0:01:01.736 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.025) 0:01:01.762 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.026) 0:01:01.788 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.026) 0:01:01.814 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.027) 0:01:01.842 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.026) 0:01:01.868 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.051) 0:01:01.920 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:25:09 +0000 (0:00:00.049) 0:01:01.970 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.039) 0:01:02.009 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.036) 0:01:02.046 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.038) 0:01:02.084 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.070) 0:01:02.155 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.085) 0:01:02.240 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.039) 0:01:02.279 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.039) 0:01:02.319 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.039) 0:01:02.359 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.039) 0:01:02.399 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.039) 0:01:02.438 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.039) 0:01:02.478 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:25:10 +0000 (0:00:00.316) 0:01:02.794 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.325) 0:01:03.120 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.050) 0:01:03.170 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.036) 0:01:03.207 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.038) 0:01:03.246 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.040) 0:01:03.286 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.040) 0:01:03.327 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.045) 0:01:03.373 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.049) 0:01:03.422 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.043) 0:01:03.466 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.038) 0:01:03.504 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.056) 0:01:03.561 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test2" ], "delta": "0:00:00.025319", "end": "2022-07-21 11:25:11.616171", "rc": 0, "start": "2022-07-21 11:25:11.590852" } STDOUT: LVM2_LV_NAME=test2 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.347) 0:01:03.908 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:25:11 +0000 (0:00:00.054) 0:01:03.963 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.054) 0:01:04.017 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.040) 0:01:04.058 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.041) 0:01:04.100 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.040) 0:01:04.140 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.039) 0:01:04.180 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.034) 0:01:04.214 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.047) 0:01:04.262 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.080) 0:01:04.342 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test3" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.077) 0:01:04.420 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test3", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test3", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "3f441d29-fa52-465c-934e-baef0ec1af88" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test3", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test3", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "3f441d29-fa52-465c-934e-baef0ec1af88" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.056) 0:01:04.477 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.084) 0:01:04.561 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.084) 0:01:04.646 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.100) 0:01:04.747 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.024) 0:01:04.772 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.023) 0:01:04.795 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.023) 0:01:04.819 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.031) 0:01:04.851 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test3 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test3 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test3 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.060) 0:01:04.911 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:25:12 +0000 (0:00:00.048) 0:01:04.959 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:25:13 +0000 (0:00:00.048) 0:01:05.008 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:25:13 +0000 (0:00:00.035) 0:01:05.043 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:25:13 +0000 (0:00:00.036) 0:01:05.079 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:25:13 +0000 (0:00:00.042) 0:01:05.121 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:25:13 +0000 (0:00:00.041) 0:01:05.163 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658417068.4414525, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658417068.4414525, "dev": 5, "device_type": 64512, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 32309, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658417068.4414525, "nlink": 1, "path": "/dev/mapper/foo-test3", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:25:13 +0000 (0:00:00.317) 0:01:05.480 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:25:13 +0000 (0:00:00.041) 0:01:05.522 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:25:13 +0000 (0:00:00.041) 0:01:05.563 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:25:13 +0000 (0:00:00.040) 0:01:05.603 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:25:13 +0000 (0:00:00.024) 0:01:05.627 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:25:13 +0000 (0:00:00.039) 0:01:05.666 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:25:13 +0000 (0:00:00.025) 0:01:05.692 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.522) 0:01:06.214 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.025) 0:01:06.240 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.023) 0:01:06.263 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.050) 0:01:06.314 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.023) 0:01:06.337 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.024) 0:01:06.361 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.023) 0:01:06.384 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.025) 0:01:06.410 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.025) 0:01:06.435 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.055) 0:01:06.490 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.084) 0:01:06.575 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.039) 0:01:06.614 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.045) 0:01:06.659 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.042) 0:01:06.701 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.091) 0:01:06.793 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.041) 0:01:06.834 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.084) 0:01:06.919 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:25:14 +0000 (0:00:00.036) 0:01:06.955 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:25:15 +0000 (0:00:00.037) 0:01:06.993 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:25:15 +0000 (0:00:00.037) 0:01:07.031 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:25:15 +0000 (0:00:00.038) 0:01:07.069 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:25:15 +0000 (0:00:00.038) 0:01:07.107 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:25:15 +0000 (0:00:00.314) 0:01:07.422 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:25:15 +0000 (0:00:00.312) 0:01:07.734 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:25:15 +0000 (0:00:00.049) 0:01:07.784 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:25:15 +0000 (0:00:00.036) 0:01:07.820 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:25:15 +0000 (0:00:00.036) 0:01:07.856 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:25:15 +0000 (0:00:00.035) 0:01:07.892 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:25:15 +0000 (0:00:00.039) 0:01:07.931 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:25:15 +0000 (0:00:00.037) 0:01:07.968 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:25:16 +0000 (0:00:00.035) 0:01:08.004 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:25:16 +0000 (0:00:00.035) 0:01:08.040 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:25:16 +0000 (0:00:00.037) 0:01:08.077 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:25:16 +0000 (0:00:00.050) 0:01:08.128 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test3" ], "delta": "0:00:00.024121", "end": "2022-07-21 11:25:16.170318", "rc": 0, "start": "2022-07-21 11:25:16.146197" } STDOUT: LVM2_LV_NAME=test3 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:25:16 +0000 (0:00:00.334) 0:01:08.462 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:25:16 +0000 (0:00:00.063) 0:01:08.525 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:25:16 +0000 (0:00:00.053) 0:01:08.579 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:25:16 +0000 (0:00:00.036) 0:01:08.616 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:25:16 +0000 (0:00:00.039) 0:01:08.656 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:25:16 +0000 (0:00:00.039) 0:01:08.695 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:25:16 +0000 (0:00:00.035) 0:01:08.730 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43 Thursday 21 July 2022 15:25:16 +0000 (0:00:00.036) 0:01:08.767 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53 Thursday 21 July 2022 15:25:16 +0000 (0:00:00.021) 0:01:08.789 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Remove two of the LVs] *************************************************** task path: /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes.yml:66 Thursday 21 July 2022 15:25:16 +0000 (0:00:00.035) 0:01:08.824 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 15:25:16 +0000 (0:00:00.052) 0:01:08.877 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 15:25:16 +0000 (0:00:00.071) 0:01:08.948 ********* ok: [/cache/rhel-7.qcow2] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 15:25:17 +0000 (0:00:00.422) 0:01:09.371 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 15:25:17 +0000 (0:00:00.063) 0:01:09.434 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 15:25:17 +0000 (0:00:00.041) 0:01:09.475 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 15:25:17 +0000 (0:00:00.035) 0:01:09.511 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 15:25:17 +0000 (0:00:00.047) 0:01:09.559 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 15:25:17 +0000 (0:00:00.020) 0:01:09.579 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 15:25:17 +0000 (0:00:00.030) 0:01:09.610 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "sdb" ], "name": "foo", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "3g" }, { "mount_point": "/opt/test2", "name": "test2", "size": "3g", "state": "absent" }, { "mount_point": "/opt/test3", "name": "test3", "size": "3g", "state": "absent" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 15:25:17 +0000 (0:00:00.036) 0:01:09.647 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 15:25:17 +0000 (0:00:00.038) 0:01:09.686 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 15:25:17 +0000 (0:00:00.029) 0:01:09.715 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 15:25:17 +0000 (0:00:00.039) 0:01:09.754 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 15:25:17 +0000 (0:00:00.038) 0:01:09.793 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 15:25:17 +0000 (0:00:00.042) 0:01:09.835 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 15:25:17 +0000 (0:00:00.057) 0:01:09.893 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 15:25:17 +0000 (0:00:00.024) 0:01:09.918 ********* changed: [/cache/rhel-7.qcow2] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test3", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test3", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/foo-test2", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test2", "fs_type": null } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/mapper/foo-test1", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "absent" }, { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "absent" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 15:25:20 +0000 (0:00:02.400) 0:01:12.318 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 15:25:20 +0000 (0:00:00.049) 0:01:12.368 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 15:25:20 +0000 (0:00:00.024) 0:01:12.392 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test3", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test3", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/foo-test2", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test2", "fs_type": null } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/mapper/foo-test1", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "absent" }, { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "absent" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 15:25:20 +0000 (0:00:00.075) 0:01:12.468 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 15:25:20 +0000 (0:00:00.041) 0:01:12.510 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 15:25:20 +0000 (0:00:00.069) 0:01:12.579 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test3', 'state': 'absent', 'path': '/opt/test3', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "absent" }, "name": "/opt/test3", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test3" } redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test2', 'state': 'absent', 'path': '/opt/test2', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "absent" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 15:25:21 +0000 (0:00:00.682) 0:01:13.262 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 15:25:21 +0000 (0:00:00.503) 0:01:13.765 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 15:25:22 +0000 (0:00:00.347) 0:01:14.112 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 15:25:22 +0000 (0:00:00.461) 0:01:14.574 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658417058.2394526, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658201031.524, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 70, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658200515.884, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744071677828413", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 15:25:22 +0000 (0:00:00.324) 0:01:14.898 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 15:25:22 +0000 (0:00:00.024) 0:01:14.923 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes.yml:86 Thursday 21 July 2022 15:25:23 +0000 (0:00:00.859) 0:01:15.782 ********* included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1 Thursday 21 July 2022 15:25:23 +0000 (0:00:00.046) 0:01:15.828 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6 Thursday 21 July 2022 15:25:23 +0000 (0:00:00.056) 0:01:15.885 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14 Thursday 21 July 2022 15:25:23 +0000 (0:00:00.038) 0:01:15.923 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "3G", "type": "lvm", "uuid": "a47eb0d6-9f88-47db-870b-ef3b58e60afc" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "c1Mk0I-V3FP-Vj0J-xjuH-291X-Hwf5-bK1aAM" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-23-55-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19 Thursday 21 July 2022 15:25:24 +0000 (0:00:00.315) 0:01:16.239 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003709", "end": "2022-07-21 11:25:24.262306", "rc": 0, "start": "2022-07-21 11:25:24.258597" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24 Thursday 21 July 2022 15:25:24 +0000 (0:00:00.314) 0:01:16.553 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003383", "end": "2022-07-21 11:25:24.570287", "failed_when_result": false, "rc": 0, "start": "2022-07-21 11:25:24.566904" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33 Thursday 21 July 2022 15:25:24 +0000 (0:00:00.306) 0:01:16.860 ********* included: /tmp/tmpaxjje44y/tests/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, {'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, {'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:5 Thursday 21 July 2022 15:25:24 +0000 (0:00:00.066) 0:01:16.927 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:18 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.082) 0:01:17.010 ********* included: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members) included: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.087) 0:01:17.097 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sdb" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.056) 0:01:17.154 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb", "pv": "/dev/sdb" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.311) 0:01:17.466 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.051) 0:01:17.517 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sdb" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.052) 0:01:17.569 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.052) 0:01:17.621 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.040) 0:01:17.661 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.052) 0:01:17.714 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.025) 0:01:17.740 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.051) 0:01:17.791 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-md.yml for /cache/rhel-7.qcow2 TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:6 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.049) 0:01:17.840 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:12 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.027) 0:01:17.868 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:16 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.025) 0:01:17.893 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:20 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.026) 0:01:17.920 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:24 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.024) 0:01:17.944 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:30 Thursday 21 July 2022 15:25:25 +0000 (0:00:00.025) 0:01:17.969 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:36 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.025) 0:01:17.995 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:44 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.025) 0:01:18.021 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.037) 0:01:18.058 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.044) 0:01:18.103 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.054) 0:01:18.158 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.027) 0:01:18.186 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.028) 0:01:18.214 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.028) 0:01:18.243 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.027) 0:01:18.270 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.028) 0:01:18.299 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.026) 0:01:18.325 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.026) 0:01:18.352 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.025) 0:01:18.378 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.028) 0:01:18.406 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml for /cache/rhel-7.qcow2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.044) 0:01:18.450 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.089) 0:01:18.540 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.025) 0:01:18.565 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.024) 0:01:18.590 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.024) 0:01:18.614 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.023) 0:01:18.637 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.022) 0:01:18.660 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.024) 0:01:18.684 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.026) 0:01:18.710 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.023) 0:01:18.734 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.025) 0:01:18.759 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.023) 0:01:18.783 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.026) 0:01:18.809 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.025) 0:01:18.835 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2 TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.048) 0:01:18.883 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.063) 0:01:18.947 ********* skipping: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "_storage_test_pool_member_path": "/dev/sdb", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 15:25:26 +0000 (0:00:00.036) 0:01:18.984 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/sdb) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.044) 0:01:19.028 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.062) 0:01:19.091 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.051) 0:01:19.142 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.037) 0:01:19.180 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.041) 0:01:19.221 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.038) 0:01:19.260 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.034) 0:01:19.294 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.037) 0:01:19.331 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.060) 0:01:19.392 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.066) 0:01:19.458 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.027) 0:01:19.485 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.026) 0:01:19.512 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.025) 0:01:19.537 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.029) 0:01:19.567 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.027) 0:01:19.594 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.025) 0:01:19.619 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.024) 0:01:19.643 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.039) 0:01:19.683 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.026) 0:01:19.709 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.026) 0:01:19.735 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.025) 0:01:19.761 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.024) 0:01:19.785 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.062) 0:01:19.848 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.025) 0:01:19.873 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.024) 0:01:19.898 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.038) 0:01:19.936 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.027) 0:01:19.963 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:25:27 +0000 (0:00:00.025) 0:01:19.989 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.023) 0:01:20.012 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.025) 0:01:20.038 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.028) 0:01:20.066 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.026) 0:01:20.092 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.025) 0:01:20.118 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.034) 0:01:20.153 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.035) 0:01:20.188 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.050) 0:01:20.239 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.054) 0:01:20.293 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.081) 0:01:20.375 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.042) 0:01:20.418 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "a47eb0d6-9f88-47db-870b-ef3b58e60afc" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "a47eb0d6-9f88-47db-870b-ef3b58e60afc" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.060) 0:01:20.479 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.056) 0:01:20.535 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.052) 0:01:20.587 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.054) 0:01:20.641 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.027) 0:01:20.669 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.024) 0:01:20.694 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.026) 0:01:20.720 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.041) 0:01:20.762 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.065) 0:01:20.827 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.051) 0:01:20.879 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.052) 0:01:20.931 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:25:28 +0000 (0:00:00.040) 0:01:20.971 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:25:29 +0000 (0:00:00.036) 0:01:21.008 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:25:29 +0000 (0:00:00.040) 0:01:21.048 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:25:29 +0000 (0:00:00.041) 0:01:21.090 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658417068.8004527, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658417068.8004527, "dev": 5, "device_type": 64514, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 32400, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658417068.8004527, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:25:29 +0000 (0:00:00.330) 0:01:21.421 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:25:29 +0000 (0:00:00.085) 0:01:21.506 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:25:29 +0000 (0:00:00.045) 0:01:21.552 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:25:29 +0000 (0:00:00.038) 0:01:21.590 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:25:29 +0000 (0:00:00.025) 0:01:21.615 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:25:29 +0000 (0:00:00.036) 0:01:21.652 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:25:29 +0000 (0:00:00.022) 0:01:21.675 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.537) 0:01:22.213 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.024) 0:01:22.237 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.022) 0:01:22.260 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.049) 0:01:22.309 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.022) 0:01:22.332 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.024) 0:01:22.356 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.027) 0:01:22.384 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.026) 0:01:22.410 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.025) 0:01:22.436 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.054) 0:01:22.490 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.052) 0:01:22.542 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.036) 0:01:22.579 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.038) 0:01:22.617 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.038) 0:01:22.656 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.036) 0:01:22.693 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.040) 0:01:22.733 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.040) 0:01:22.773 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.038) 0:01:22.812 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.041) 0:01:22.853 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.038) 0:01:22.892 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.035) 0:01:22.928 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:25:30 +0000 (0:00:00.035) 0:01:22.963 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:25:31 +0000 (0:00:00.316) 0:01:23.279 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:25:31 +0000 (0:00:00.350) 0:01:23.630 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:25:31 +0000 (0:00:00.104) 0:01:23.734 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:25:31 +0000 (0:00:00.037) 0:01:23.771 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:25:31 +0000 (0:00:00.037) 0:01:23.809 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:25:31 +0000 (0:00:00.038) 0:01:23.847 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:25:31 +0000 (0:00:00.036) 0:01:23.884 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:25:31 +0000 (0:00:00.039) 0:01:23.923 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:25:31 +0000 (0:00:00.039) 0:01:23.963 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:25:32 +0000 (0:00:00.075) 0:01:24.039 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:25:32 +0000 (0:00:00.077) 0:01:24.117 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:25:32 +0000 (0:00:00.095) 0:01:24.212 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.027355", "end": "2022-07-21 11:25:32.306246", "rc": 0, "start": "2022-07-21 11:25:32.278891" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:25:32 +0000 (0:00:00.384) 0:01:24.597 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:25:32 +0000 (0:00:00.050) 0:01:24.648 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:25:32 +0000 (0:00:00.053) 0:01:24.701 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:25:32 +0000 (0:00:00.037) 0:01:24.738 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:25:32 +0000 (0:00:00.039) 0:01:24.777 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:25:32 +0000 (0:00:00.037) 0:01:24.815 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:25:32 +0000 (0:00:00.036) 0:01:24.852 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:25:32 +0000 (0:00:00.033) 0:01:24.885 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:25:32 +0000 (0:00:00.049) 0:01:24.935 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.080) 0:01:25.016 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.042) 0:01:25.058 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.058) 0:01:25.116 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.025) 0:01:25.142 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.051) 0:01:25.194 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.037) 0:01:25.231 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.025) 0:01:25.257 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.026) 0:01:25.284 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.025) 0:01:25.309 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.034) 0:01:25.343 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.061) 0:01:25.405 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.024) 0:01:25.429 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.050) 0:01:25.480 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.038) 0:01:25.519 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.035) 0:01:25.555 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.026) 0:01:25.582 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.024) 0:01:25.606 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.312) 0:01:25.918 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:25:33 +0000 (0:00:00.045) 0:01:25.964 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:25:34 +0000 (0:00:00.026) 0:01:25.991 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:25:34 +0000 (0:00:00.039) 0:01:26.031 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:25:34 +0000 (0:00:00.025) 0:01:26.056 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:25:34 +0000 (0:00:00.025) 0:01:26.082 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:25:34 +0000 (0:00:00.023) 0:01:26.106 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:25:34 +0000 (0:00:00.613) 0:01:26.720 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:25:34 +0000 (0:00:00.027) 0:01:26.747 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:25:34 +0000 (0:00:00.024) 0:01:26.772 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:25:34 +0000 (0:00:00.023) 0:01:26.795 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:25:34 +0000 (0:00:00.023) 0:01:26.819 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:25:34 +0000 (0:00:00.023) 0:01:26.842 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:25:34 +0000 (0:00:00.023) 0:01:26.865 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:25:34 +0000 (0:00:00.022) 0:01:26.888 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:25:34 +0000 (0:00:00.022) 0:01:26.911 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:25:34 +0000 (0:00:00.051) 0:01:26.962 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.048) 0:01:27.011 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.035) 0:01:27.046 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.036) 0:01:27.082 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.038) 0:01:27.121 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.032) 0:01:27.154 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.038) 0:01:27.193 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.036) 0:01:27.230 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.041) 0:01:27.271 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.039) 0:01:27.311 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.039) 0:01:27.350 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.038) 0:01:27.388 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.041) 0:01:27.430 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.026) 0:01:27.457 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.039) 0:01:27.497 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.037) 0:01:27.535 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.038) 0:01:27.573 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.034) 0:01:27.608 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.035) 0:01:27.643 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.035) 0:01:27.678 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.038) 0:01:27.717 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.038) 0:01:27.756 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.038) 0:01:27.795 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.037) 0:01:27.833 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.026) 0:01:27.859 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.026) 0:01:27.885 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.026) 0:01:27.911 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.025) 0:01:27.937 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:25:35 +0000 (0:00:00.027) 0:01:27.965 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.058) 0:01:28.023 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.026) 0:01:28.050 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.026) 0:01:28.076 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.037) 0:01:28.114 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.061) 0:01:28.175 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.078) 0:01:28.253 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test3" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.040) 0:01:28.294 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.055) 0:01:28.350 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.027) 0:01:28.377 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.053) 0:01:28.430 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.039) 0:01:28.470 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.026) 0:01:28.497 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.026) 0:01:28.523 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.025) 0:01:28.549 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.038) 0:01:28.587 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.065) 0:01:28.653 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.026) 0:01:28.679 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.049) 0:01:28.729 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.047) 0:01:28.776 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.040) 0:01:28.817 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.025) 0:01:28.842 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:25:36 +0000 (0:00:00.026) 0:01:28.869 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:25:37 +0000 (0:00:00.324) 0:01:29.193 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:25:37 +0000 (0:00:00.039) 0:01:29.233 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:25:37 +0000 (0:00:00.023) 0:01:29.256 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:25:37 +0000 (0:00:00.036) 0:01:29.292 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:25:37 +0000 (0:00:00.023) 0:01:29.315 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:25:37 +0000 (0:00:00.023) 0:01:29.339 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:25:37 +0000 (0:00:00.023) 0:01:29.362 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:25:37 +0000 (0:00:00.531) 0:01:29.894 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:25:37 +0000 (0:00:00.026) 0:01:29.920 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:25:37 +0000 (0:00:00.025) 0:01:29.945 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:25:37 +0000 (0:00:00.022) 0:01:29.968 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.024) 0:01:29.993 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.023) 0:01:30.016 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.025) 0:01:30.042 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.023) 0:01:30.066 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.065) 0:01:30.131 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.050) 0:01:30.182 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.053) 0:01:30.235 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.040) 0:01:30.276 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.040) 0:01:30.317 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.037) 0:01:30.354 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.035) 0:01:30.390 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.040) 0:01:30.430 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.047) 0:01:30.478 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.040) 0:01:30.519 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.039) 0:01:30.559 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.042) 0:01:30.602 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.039) 0:01:30.641 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.036) 0:01:30.677 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.024) 0:01:30.701 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.036) 0:01:30.738 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.034) 0:01:30.772 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.034) 0:01:30.807 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.034) 0:01:30.842 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.036) 0:01:30.878 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.039) 0:01:30.917 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:25:38 +0000 (0:00:00.038) 0:01:30.956 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.038) 0:01:30.995 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.041) 0:01:31.037 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.037) 0:01:31.075 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.024) 0:01:31.100 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.026) 0:01:31.126 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.026) 0:01:31.153 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.026) 0:01:31.179 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.025) 0:01:31.205 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.025) 0:01:31.231 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.026) 0:01:31.257 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.023) 0:01:31.281 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.037) 0:01:31.318 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.023) 0:01:31.341 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Re-run the previous role invocation to ensure idempotence] *************** task path: /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes.yml:88 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.044) 0:01:31.386 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.055) 0:01:31.441 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.035) 0:01:31.477 ********* ok: [/cache/rhel-7.qcow2] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 15:25:39 +0000 (0:00:00.452) 0:01:31.930 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 15:25:40 +0000 (0:00:00.060) 0:01:31.990 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 15:25:40 +0000 (0:00:00.031) 0:01:32.021 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 15:25:40 +0000 (0:00:00.032) 0:01:32.054 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 15:25:40 +0000 (0:00:00.046) 0:01:32.100 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 15:25:40 +0000 (0:00:00.020) 0:01:32.121 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 15:25:40 +0000 (0:00:00.027) 0:01:32.148 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "sdb" ], "name": "foo", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "3g" }, { "mount_point": "/opt/test2", "name": "test2", "size": "3g", "state": "absent" }, { "mount_point": "/opt/test3", "name": "test3", "size": "3g", "state": "absent" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 15:25:40 +0000 (0:00:00.038) 0:01:32.187 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 15:25:40 +0000 (0:00:00.037) 0:01:32.225 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 15:25:40 +0000 (0:00:00.030) 0:01:32.255 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 15:25:40 +0000 (0:00:00.029) 0:01:32.285 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 15:25:40 +0000 (0:00:00.031) 0:01:32.316 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 15:25:40 +0000 (0:00:00.029) 0:01:32.346 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 15:25:40 +0000 (0:00:00.059) 0:01:32.405 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 15:25:40 +0000 (0:00:00.021) 0:01:32.426 ********* ok: [/cache/rhel-7.qcow2] => { "actions": [], "changed": false, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/mapper/foo-test1", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "", "_mount_id": "", "_raw_device": "", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "", "_mount_id": "", "_raw_device": "", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 15:25:41 +0000 (0:00:01.208) 0:01:33.635 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 15:25:41 +0000 (0:00:00.035) 0:01:33.671 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 15:25:41 +0000 (0:00:00.024) 0:01:33.696 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/mapper/foo-test1", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "", "_mount_id": "", "_raw_device": "", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "", "_mount_id": "", "_raw_device": "", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 15:25:41 +0000 (0:00:00.043) 0:01:33.740 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "", "_mount_id": "", "_raw_device": "", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "", "_mount_id": "", "_raw_device": "", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 15:25:41 +0000 (0:00:00.045) 0:01:33.785 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 15:25:41 +0000 (0:00:00.038) 0:01:33.824 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 15:25:41 +0000 (0:00:00.040) 0:01:33.864 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 15:25:42 +0000 (0:00:00.502) 0:01:34.366 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 15:25:42 +0000 (0:00:00.381) 0:01:34.748 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 15:25:43 +0000 (0:00:00.475) 0:01:35.223 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658417058.2394526, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658201031.524, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 70, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658200515.884, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744071677828413", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 15:25:43 +0000 (0:00:00.357) 0:01:35.581 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 15:25:43 +0000 (0:00:00.022) 0:01:35.603 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes.yml:108 Thursday 21 July 2022 15:25:44 +0000 (0:00:00.847) 0:01:36.451 ********* included: /tmp/tmpaxjje44y/tests/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:1 Thursday 21 July 2022 15:25:44 +0000 (0:00:00.047) 0:01:36.498 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "", "_mount_id": "", "_raw_device": "", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "", "_mount_id": "", "_raw_device": "", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:6 Thursday 21 July 2022 15:25:44 +0000 (0:00:00.053) 0:01:36.552 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:14 Thursday 21 July 2022 15:25:44 +0000 (0:00:00.036) 0:01:36.588 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "3G", "type": "lvm", "uuid": "a47eb0d6-9f88-47db-870b-ef3b58e60afc" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "c1Mk0I-V3FP-Vj0J-xjuH-291X-Hwf5-bK1aAM" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-15-23-55-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:19 Thursday 21 July 2022 15:25:44 +0000 (0:00:00.319) 0:01:36.908 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.004313", "end": "2022-07-21 11:25:44.935328", "rc": 0, "start": "2022-07-21 11:25:44.931015" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:24 Thursday 21 July 2022 15:25:45 +0000 (0:00:00.321) 0:01:37.230 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003951", "end": "2022-07-21 11:25:45.258351", "failed_when_result": false, "rc": 0, "start": "2022-07-21 11:25:45.254400" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:33 Thursday 21 July 2022 15:25:45 +0000 (0:00:00.321) 0:01:37.551 ********* included: /tmp/tmpaxjje44y/tests/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, {'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, {'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:5 Thursday 21 July 2022 15:25:45 +0000 (0:00:00.063) 0:01:37.615 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool.yml:18 Thursday 21 July 2022 15:25:45 +0000 (0:00:00.035) 0:01:37.651 ********* included: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members) included: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 15:25:45 +0000 (0:00:00.047) 0:01:37.698 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sdb" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 15:25:45 +0000 (0:00:00.090) 0:01:37.789 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb", "pv": "/dev/sdb" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.332) 0:01:38.122 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.050) 0:01:38.172 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sdb" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.088) 0:01:38.260 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.131) 0:01:38.392 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.040) 0:01:38.432 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.054) 0:01:38.487 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.025) 0:01:38.513 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.046) 0:01:38.559 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-md.yml for /cache/rhel-7.qcow2 TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:6 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.045) 0:01:38.604 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:12 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.028) 0:01:38.633 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:16 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.027) 0:01:38.660 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:20 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.027) 0:01:38.687 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:24 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.027) 0:01:38.715 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:30 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.032) 0:01:38.748 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:36 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.031) 0:01:38.779 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-md.yml:44 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.027) 0:01:38.807 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.040) 0:01:38.847 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.048) 0:01:38.895 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.057) 0:01:38.953 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:25:46 +0000 (0:00:00.032) 0:01:38.985 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.030) 0:01:39.016 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.029) 0:01:39.046 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.029) 0:01:39.076 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.031) 0:01:39.108 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.029) 0:01:39.137 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.029) 0:01:39.167 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.029) 0:01:39.197 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.030) 0:01:39.227 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml for /cache/rhel-7.qcow2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.046) 0:01:39.274 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.058) 0:01:39.333 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.026) 0:01:39.359 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.026) 0:01:39.385 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.027) 0:01:39.412 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.025) 0:01:39.438 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.025) 0:01:39.463 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.025) 0:01:39.489 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.027) 0:01:39.516 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.025) 0:01:39.541 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.025) 0:01:39.567 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.024) 0:01:39.592 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.027) 0:01:39.619 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.068) 0:01:39.688 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2 TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.048) 0:01:39.736 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.054) 0:01:39.791 ********* skipping: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "_storage_test_pool_member_path": "/dev/sdb", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.029) 0:01:39.821 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/sdb) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.044) 0:01:39.865 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.053) 0:01:39.919 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 15:25:47 +0000 (0:00:00.051) 0:01:39.970 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.038) 0:01:40.009 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.038) 0:01:40.047 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.043) 0:01:40.090 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.035) 0:01:40.126 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.037) 0:01:40.163 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.048) 0:01:40.211 ********* included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.057) 0:01:40.269 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.028) 0:01:40.298 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.028) 0:01:40.327 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.026) 0:01:40.353 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.025) 0:01:40.379 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.027) 0:01:40.406 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.026) 0:01:40.433 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.026) 0:01:40.460 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.038) 0:01:40.499 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.028) 0:01:40.527 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.026) 0:01:40.554 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.026) 0:01:40.580 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.026) 0:01:40.606 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.027) 0:01:40.634 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.026) 0:01:40.661 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.026) 0:01:40.688 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.038) 0:01:40.726 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.028) 0:01:40.754 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.027) 0:01:40.782 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.026) 0:01:40.808 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.026) 0:01:40.835 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.027) 0:01:40.862 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.025) 0:01:40.888 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 15:25:48 +0000 (0:00:00.026) 0:01:40.914 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.077) 0:01:40.992 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.081) 0:01:41.073 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmpaxjje44y/tests/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.054) 0:01:41.128 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.051) 0:01:41.180 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.081) 0:01:41.261 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.044) 0:01:41.305 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "a47eb0d6-9f88-47db-870b-ef3b58e60afc" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "a47eb0d6-9f88-47db-870b-ef3b58e60afc" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.059) 0:01:41.365 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.053) 0:01:41.418 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.052) 0:01:41.470 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.054) 0:01:41.525 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.027) 0:01:41.553 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.026) 0:01:41.579 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.026) 0:01:41.606 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.038) 0:01:41.644 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.064) 0:01:41.709 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.052) 0:01:41.761 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.054) 0:01:41.816 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.040) 0:01:41.856 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.037) 0:01:41.893 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.041) 0:01:41.935 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:25:49 +0000 (0:00:00.039) 0:01:41.974 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658417068.8004527, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658417068.8004527, "dev": 5, "device_type": 64514, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 32400, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658417068.8004527, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:25:50 +0000 (0:00:00.331) 0:01:42.306 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:25:50 +0000 (0:00:00.040) 0:01:42.346 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:25:50 +0000 (0:00:00.042) 0:01:42.389 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:25:50 +0000 (0:00:00.038) 0:01:42.427 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:25:50 +0000 (0:00:00.027) 0:01:42.455 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:25:50 +0000 (0:00:00.041) 0:01:42.496 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:25:50 +0000 (0:00:00.024) 0:01:42.520 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.551) 0:01:43.072 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.024) 0:01:43.096 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.033) 0:01:43.129 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.057) 0:01:43.186 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.026) 0:01:43.213 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.026) 0:01:43.239 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.029) 0:01:43.269 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.026) 0:01:43.296 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.026) 0:01:43.322 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.119) 0:01:43.441 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.107) 0:01:43.549 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.036) 0:01:43.586 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.039) 0:01:43.625 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.041) 0:01:43.667 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.034) 0:01:43.701 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.036) 0:01:43.738 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.040) 0:01:43.778 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.039) 0:01:43.818 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.037) 0:01:43.856 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.038) 0:01:43.894 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.038) 0:01:43.932 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:25:51 +0000 (0:00:00.039) 0:01:43.971 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:25:52 +0000 (0:00:00.322) 0:01:44.294 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:25:52 +0000 (0:00:00.349) 0:01:44.643 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:25:52 +0000 (0:00:00.052) 0:01:44.695 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:25:52 +0000 (0:00:00.037) 0:01:44.733 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:25:52 +0000 (0:00:00.039) 0:01:44.773 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:25:52 +0000 (0:00:00.045) 0:01:44.818 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:25:52 +0000 (0:00:00.040) 0:01:44.858 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:25:52 +0000 (0:00:00.041) 0:01:44.899 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:25:52 +0000 (0:00:00.038) 0:01:44.938 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:25:52 +0000 (0:00:00.040) 0:01:44.979 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:25:53 +0000 (0:00:00.040) 0:01:45.020 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:25:53 +0000 (0:00:00.059) 0:01:45.079 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.026527", "end": "2022-07-21 11:25:53.134172", "rc": 0, "start": "2022-07-21 11:25:53.107645" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:25:53 +0000 (0:00:00.349) 0:01:45.429 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:25:53 +0000 (0:00:00.054) 0:01:45.483 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:25:53 +0000 (0:00:00.054) 0:01:45.538 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:25:53 +0000 (0:00:00.039) 0:01:45.578 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:25:53 +0000 (0:00:00.039) 0:01:45.618 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:25:53 +0000 (0:00:00.039) 0:01:45.657 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:25:53 +0000 (0:00:00.040) 0:01:45.698 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:25:53 +0000 (0:00:00.039) 0:01:45.738 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:25:53 +0000 (0:00:00.097) 0:01:45.835 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:25:53 +0000 (0:00:00.124) 0:01:45.959 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.045) 0:01:46.005 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.056) 0:01:46.061 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.025) 0:01:46.087 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.050) 0:01:46.137 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.037) 0:01:46.175 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.026) 0:01:46.201 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.024) 0:01:46.226 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.026) 0:01:46.252 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.040) 0:01:46.292 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.059) 0:01:46.352 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.025) 0:01:46.378 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.051) 0:01:46.430 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.037) 0:01:46.468 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.035) 0:01:46.503 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.023) 0:01:46.527 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.024) 0:01:46.552 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.312) 0:01:46.864 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.039) 0:01:46.904 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.024) 0:01:46.928 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:25:54 +0000 (0:00:00.036) 0:01:46.964 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:25:55 +0000 (0:00:00.026) 0:01:46.991 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:25:55 +0000 (0:00:00.024) 0:01:47.015 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:25:55 +0000 (0:00:00.024) 0:01:47.040 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:25:55 +0000 (0:00:00.531) 0:01:47.572 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:25:55 +0000 (0:00:00.024) 0:01:47.597 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:25:55 +0000 (0:00:00.026) 0:01:47.623 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:25:55 +0000 (0:00:00.024) 0:01:47.648 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:25:55 +0000 (0:00:00.022) 0:01:47.670 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:25:55 +0000 (0:00:00.024) 0:01:47.695 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:25:55 +0000 (0:00:00.024) 0:01:47.719 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:25:55 +0000 (0:00:00.022) 0:01:47.742 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:25:55 +0000 (0:00:00.022) 0:01:47.765 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:25:55 +0000 (0:00:00.084) 0:01:47.850 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:25:55 +0000 (0:00:00.101) 0:01:47.951 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:25:55 +0000 (0:00:00.037) 0:01:47.988 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.038) 0:01:48.027 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.039) 0:01:48.066 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.035) 0:01:48.102 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.038) 0:01:48.141 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.037) 0:01:48.178 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.036) 0:01:48.214 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.037) 0:01:48.252 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.034) 0:01:48.287 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.034) 0:01:48.321 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.035) 0:01:48.357 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.023) 0:01:48.380 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.038) 0:01:48.418 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.038) 0:01:48.457 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.039) 0:01:48.496 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.039) 0:01:48.535 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.037) 0:01:48.573 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.037) 0:01:48.611 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.039) 0:01:48.650 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.038) 0:01:48.688 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.036) 0:01:48.725 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.040) 0:01:48.765 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.026) 0:01:48.791 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.025) 0:01:48.817 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.024) 0:01:48.842 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.026) 0:01:48.868 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.025) 0:01:48.893 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.022) 0:01:48.916 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.022) 0:01:48.939 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:25:56 +0000 (0:00:00.025) 0:01:48.964 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:2 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.039) 0:01:49.004 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:10 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.048) 0:01:49.053 ********* included: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.081) 0:01:49.134 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.042) 0:01:49.177 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.054) 0:01:49.232 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.022) 0:01:49.254 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.078) 0:01:49.333 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.035) 0:01:49.368 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.057) 0:01:49.426 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.024) 0:01:49.450 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.025) 0:01:49.475 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.035) 0:01:49.511 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.066) 0:01:49.577 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.025) 0:01:49.603 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.051) 0:01:49.654 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.039) 0:01:49.694 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.035) 0:01:49.730 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.024) 0:01:49.754 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 15:25:57 +0000 (0:00:00.024) 0:01:49.779 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 15:25:58 +0000 (0:00:00.304) 0:01:50.083 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 15:25:58 +0000 (0:00:00.041) 0:01:50.125 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 15:25:58 +0000 (0:00:00.025) 0:01:50.150 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 15:25:58 +0000 (0:00:00.039) 0:01:50.190 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 15:25:58 +0000 (0:00:00.025) 0:01:50.215 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 15:25:58 +0000 (0:00:00.024) 0:01:50.240 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 15:25:58 +0000 (0:00:00.028) 0:01:50.268 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 15:25:58 +0000 (0:00:00.520) 0:01:50.789 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 15:25:58 +0000 (0:00:00.027) 0:01:50.816 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 15:25:58 +0000 (0:00:00.024) 0:01:50.840 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 15:25:58 +0000 (0:00:00.026) 0:01:50.867 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 15:25:58 +0000 (0:00:00.028) 0:01:50.895 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 15:25:58 +0000 (0:00:00.023) 0:01:50.919 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 15:25:58 +0000 (0:00:00.023) 0:01:50.943 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 15:25:58 +0000 (0:00:00.024) 0:01:50.967 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.022) 0:01:50.990 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.050) 0:01:51.041 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.050) 0:01:51.091 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.041) 0:01:51.133 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.037) 0:01:51.170 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.039) 0:01:51.210 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.036) 0:01:51.246 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.041) 0:01:51.287 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.036) 0:01:51.323 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.035) 0:01:51.359 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.038) 0:01:51.398 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.038) 0:01:51.437 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.035) 0:01:51.472 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.039) 0:01:51.511 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.025) 0:01:51.537 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.041) 0:01:51.579 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.039) 0:01:51.619 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.091) 0:01:51.711 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.040) 0:01:51.752 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.039) 0:01:51.791 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.035) 0:01:51.827 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.035) 0:01:51.863 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.078) 0:01:51.941 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [debug] ******************************************************************* task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 15:25:59 +0000 (0:00:00.035) 0:01:51.977 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.037) 0:01:52.015 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.026) 0:01:52.041 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.026) 0:01:52.067 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check segment type] ****************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.026) 0:01:52.094 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.023) 0:01:52.118 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.025) 0:01:52.143 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.026) 0:01:52.169 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.024) 0:01:52.194 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpaxjje44y/tests/test-verify-volume.yml:16 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.024) 0:01:52.219 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:43 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.034) 0:01:52.253 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpaxjje44y/tests/verify-role-results.yml:53 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.022) 0:01:52.275 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Clean up] **************************************************************** task path: /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes.yml:110 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.034) 0:01:52.310 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.064) 0:01:52.374 ********* included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.034) 0:01:52.409 ********* ok: [/cache/rhel-7.qcow2] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.438) 0:01:52.847 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.060) 0:01:52.907 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.033) 0:01:52.941 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 15:26:00 +0000 (0:00:00.035) 0:01:52.977 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 15:26:01 +0000 (0:00:00.046) 0:01:53.023 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 15:26:01 +0000 (0:00:00.022) 0:01:53.045 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 15:26:01 +0000 (0:00:00.028) 0:01:53.074 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "sdb" ], "name": "foo", "state": "absent", "volumes": [] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 15:26:01 +0000 (0:00:00.036) 0:01:53.110 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 15:26:01 +0000 (0:00:00.035) 0:01:53.146 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 15:26:01 +0000 (0:00:00.028) 0:01:53.175 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 15:26:01 +0000 (0:00:00.029) 0:01:53.204 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 15:26:01 +0000 (0:00:00.026) 0:01:53.231 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 15:26:01 +0000 (0:00:00.026) 0:01:53.258 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 15:26:01 +0000 (0:00:00.055) 0:01:53.314 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 15:26:01 +0000 (0:00:00.021) 0:01:53.335 ********* changed: [/cache/rhel-7.qcow2] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdb", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "xfsprogs" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 15:26:02 +0000 (0:00:01.616) 0:01:54.952 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 15:26:02 +0000 (0:00:00.036) 0:01:54.988 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 15:26:03 +0000 (0:00:00.021) 0:01:55.010 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdb", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "xfsprogs" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 15:26:03 +0000 (0:00:00.036) 0:01:55.046 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 15:26:03 +0000 (0:00:00.039) 0:01:55.086 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 15:26:03 +0000 (0:00:00.038) 0:01:55.124 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 15:26:03 +0000 (0:00:00.349) 0:01:55.473 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 15:26:03 +0000 (0:00:00.470) 0:01:55.944 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 15:26:03 +0000 (0:00:00.040) 0:01:55.984 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 15:26:04 +0000 (0:00:00.466) 0:01:56.451 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658417058.2394526, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658201031.524, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 70, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658200515.884, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744071677828413", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 15:26:04 +0000 (0:00:00.326) 0:01:56.777 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 15:26:04 +0000 (0:00:00.022) 0:01:56.800 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* /cache/rhel-7.qcow2 : ok=748 changed=9 unreachable=0 failed=0 skipped=735 rescued=0 ignored=0 Thursday 21 July 2022 15:26:05 +0000 (0:00:00.851) 0:01:57.652 ********* =============================================================================== linux-system-roles.storage : make sure blivet is available -------------- 8.81s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 2.40s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.89s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.72s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.62s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Ensure cryptsetup is present -------------------------------------------- 1.53s /tmp/tmpaxjje44y/tests/test-verify-volume-encryption.yml:10 ------------------- set up internal repositories -------------------------------------------- 1.41s /cache/rhel-7_setup.yml:5 ----------------------------------------------------- linux-system-roles.storage : manage the pools and volumes to match the specified state --- 1.21s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : get service facts -------------------------- 1.17s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 linux-system-roles.storage : set up new/current mounts ------------------ 1.11s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Gathering Facts --------------------------------------------------------- 1.01s /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes_scsi_generated.yml:3 linux-system-roles.storage : set up new/current mounts ------------------ 0.95s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 linux-system-roles.storage : Update facts ------------------------------- 0.93s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 linux-system-roles.storage : Update facts ------------------------------- 0.87s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 linux-system-roles.storage : Update facts ------------------------------- 0.86s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 linux-system-roles.storage : Update facts ------------------------------- 0.85s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 linux-system-roles.storage : Update facts ------------------------------- 0.85s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 linux-system-roles.storage : Update facts ------------------------------- 0.85s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Gathering Facts --------------------------------------------------------- 0.75s /tmp/tmpaxjje44y/tests/tests_lvm_one_disk_multiple_volumes.yml:2 -------------- linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab --- 0.71s /tmp/tmpaxjje44y/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 ansible-playbook [core 2.12.6] config file = /etc/ansible/ansible.cfg configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.9/site-packages/ansible ansible collection location = /tmp/tmp5bkr4li_ executable location = /usr/bin/ansible-playbook python version = 3.9.13 (main, May 18 2022, 00:00:00) [GCC 11.3.1 20220421 (Red Hat 11.3.1-2)] jinja version = 2.11.3 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: rhel-7_setup.yml ***************************************************** 1 plays in /cache/rhel-7_setup.yml PLAY [Setup repos] ************************************************************* META: ran handlers TASK [set up internal repositories] ******************************************** task path: /cache/rhel-7_setup.yml:5 Thursday 21 July 2022 18:33:16 +0000 (0:00:00.017) 0:00:00.017 ********* changed: [/cache/rhel-7.qcow2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [/cache/rhel-7.qcow2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [/cache/rhel-7.qcow2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [/cache/rhel-7.qcow2] => (item=None) => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } changed: [/cache/rhel-7.qcow2] => { "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true } META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* /cache/rhel-7.qcow2 : ok=1 changed=1 unreachable=0 failed=0 skipped=0 rescued=0 ignored=0 Thursday 21 July 2022 18:33:18 +0000 (0:00:01.417) 0:00:01.435 ********* =============================================================================== set up internal repositories -------------------------------------------- 1.42s /cache/rhel-7_setup.yml:5 ----------------------------------------------------- PLAYBOOK: tests_lvm_one_disk_multiple_volumes_scsi_generated.yml *************** 2 plays in /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes_scsi_generated.yml PLAY [all] ********************************************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes_scsi_generated.yml:3 Thursday 21 July 2022 18:33:18 +0000 (0:00:00.015) 0:00:01.450 ********* ok: [/cache/rhel-7.qcow2] META: ran handlers TASK [set disk interface for test] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes_scsi_generated.yml:7 Thursday 21 July 2022 18:33:19 +0000 (0:00:00.996) 0:00:02.447 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_use_interface": "scsi" }, "changed": false } META: ran handlers META: ran handlers PLAY [all] ********************************************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes.yml:2 Thursday 21 July 2022 18:33:19 +0000 (0:00:00.062) 0:00:02.510 ********* ok: [/cache/rhel-7.qcow2] META: ran handlers TASK [include_role : fedora.linux_system_roles.storage] ************************ task path: /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes.yml:11 Thursday 21 July 2022 18:33:19 +0000 (0:00:00.736) 0:00:03.246 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 18:33:19 +0000 (0:00:00.035) 0:00:03.281 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 18:33:19 +0000 (0:00:00.028) 0:00:03.310 ********* ok: [/cache/rhel-7.qcow2] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 18:33:20 +0000 (0:00:00.411) 0:00:03.722 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 18:33:20 +0000 (0:00:00.055) 0:00:03.778 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 18:33:20 +0000 (0:00:00.035) 0:00:03.813 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 18:33:20 +0000 (0:00:00.034) 0:00:03.848 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 18:33:20 +0000 (0:00:00.056) 0:00:03.904 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 18:33:20 +0000 (0:00:00.017) 0:00:03.922 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "changes": { "installed": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "rc": 0, "results": [ "Loaded plugins: search-disabled-repos\nResolving Dependencies\n--> Running transaction check\n---> Package libblockdev-crypto.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libblockdev-utils(x86-64) = 2.18-5.el7 for package: libblockdev-crypto-2.18-5.el7.x86_64\n--> Processing Dependency: libvolume_key.so.1()(64bit) for package: libblockdev-crypto-2.18-5.el7.x86_64\n--> Processing Dependency: libbd_utils.so.2()(64bit) for package: libblockdev-crypto-2.18-5.el7.x86_64\n---> Package libblockdev-dm.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libdmraid.so.1(Base)(64bit) for package: libblockdev-dm-2.18-5.el7.x86_64\n--> Processing Dependency: dmraid for package: libblockdev-dm-2.18-5.el7.x86_64\n--> Processing Dependency: libdmraid.so.1()(64bit) for package: libblockdev-dm-2.18-5.el7.x86_64\n---> Package libblockdev-lvm.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: lvm2 for package: libblockdev-lvm-2.18-5.el7.x86_64\n--> Processing Dependency: device-mapper-persistent-data for package: libblockdev-lvm-2.18-5.el7.x86_64\n---> Package libblockdev-mdraid.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: mdadm for package: libblockdev-mdraid-2.18-5.el7.x86_64\n--> Processing Dependency: libbytesize.so.1()(64bit) for package: libblockdev-mdraid-2.18-5.el7.x86_64\n---> Package libblockdev-swap.x86_64 0:2.18-5.el7 will be installed\n---> Package python-enum34.noarch 0:1.0.4-1.el7 will be installed\n---> Package python2-blivet3.noarch 1:3.1.3-3.el7 will be installed\n--> Processing Dependency: blivet3-data = 1:3.1.3-3.el7 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-bytesize >= 0.3 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-blockdev >= 2.17 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: pyparted >= 3.9 for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: python2-hawkey for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Processing Dependency: lsof for package: 1:python2-blivet3-3.1.3-3.el7.noarch\n--> Running transaction check\n---> Package blivet3-data.noarch 1:3.1.3-3.el7 will be installed\n---> Package device-mapper-persistent-data.x86_64 0:0.8.5-3.el7_9.2 will be installed\n--> Processing Dependency: libaio.so.1(LIBAIO_0.4)(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n--> Processing Dependency: libaio.so.1(LIBAIO_0.1)(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n--> Processing Dependency: libaio.so.1()(64bit) for package: device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64\n---> Package dmraid.x86_64 0:1.0.0.rc16-28.el7 will be installed\n--> Processing Dependency: libdevmapper-event.so.1.02(Base)(64bit) for package: dmraid-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: dmraid-events for package: dmraid-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: libdevmapper-event.so.1.02()(64bit) for package: dmraid-1.0.0.rc16-28.el7.x86_64\n---> Package libblockdev-utils.x86_64 0:2.18-5.el7 will be installed\n---> Package libbytesize.x86_64 0:1.2-1.el7 will be installed\n--> Processing Dependency: libmpfr.so.4()(64bit) for package: libbytesize-1.2-1.el7.x86_64\n---> Package lsof.x86_64 0:4.87-6.el7 will be installed\n---> Package lvm2.x86_64 7:2.02.187-6.el7_9.5 will be installed\n--> Processing Dependency: lvm2-libs = 7:2.02.187-6.el7_9.5 for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n--> Processing Dependency: liblvm2app.so.2.2(Base)(64bit) for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n--> Processing Dependency: liblvm2app.so.2.2()(64bit) for package: 7:lvm2-2.02.187-6.el7_9.5.x86_64\n---> Package mdadm.x86_64 0:4.1-9.el7_9 will be installed\n--> Processing Dependency: libreport-filesystem for package: mdadm-4.1-9.el7_9.x86_64\n---> Package pyparted.x86_64 1:3.9-15.el7 will be installed\n---> Package python2-blockdev.x86_64 0:2.18-5.el7 will be installed\n--> Processing Dependency: libblockdev(x86-64) = 2.18-5.el7 for package: python2-blockdev-2.18-5.el7.x86_64\n---> Package python2-bytesize.x86_64 0:1.2-1.el7 will be installed\n---> Package python2-hawkey.x86_64 0:0.22.5-2.el7_9 will be installed\n--> Processing Dependency: libdnf(x86-64) = 0.22.5-2.el7_9 for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: python2-libdnf = 0.22.5-2.el7_9 for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolv.so.0(SOLV_1.0)(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolvext.so.0(SOLV_1.0)(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libdnf.so.2()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libjson-glib-1.0.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libmodulemd.so.1()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: librepo.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: librhsm.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolv.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n--> Processing Dependency: libsolvext.so.0()(64bit) for package: python2-hawkey-0.22.5-2.el7_9.x86_64\n---> Package volume_key-libs.x86_64 0:0.3.9-9.el7 will be installed\n--> Running transaction check\n---> Package device-mapper-event-libs.x86_64 7:1.02.170-6.el7_9.5 will be installed\n---> Package dmraid-events.x86_64 0:1.0.0.rc16-28.el7 will be installed\n--> Processing Dependency: sgpio for package: dmraid-events-1.0.0.rc16-28.el7.x86_64\n--> Processing Dependency: device-mapper-event for package: dmraid-events-1.0.0.rc16-28.el7.x86_64\n---> Package json-glib.x86_64 0:1.4.2-2.el7 will be installed\n---> Package libaio.x86_64 0:0.3.109-13.el7 will be installed\n---> Package libblockdev.x86_64 0:2.18-5.el7 will be installed\n---> Package libdnf.x86_64 0:0.22.5-2.el7_9 will be installed\n---> Package libmodulemd.x86_64 0:1.6.3-1.el7 will be installed\n---> Package librepo.x86_64 0:1.8.1-8.el7_9 will be installed\n---> Package libreport-filesystem.x86_64 0:2.1.11-53.el7 will be installed\n---> Package librhsm.x86_64 0:0.0.3-3.el7_9 will be installed\n---> Package libsolv.x86_64 0:0.6.34-4.el7 will be installed\n---> Package lvm2-libs.x86_64 7:2.02.187-6.el7_9.5 will be installed\n---> Package mpfr.x86_64 0:3.1.1-4.el7 will be installed\n---> Package python2-libdnf.x86_64 0:0.22.5-2.el7_9 will be installed\n--> Running transaction check\n---> Package device-mapper-event.x86_64 7:1.02.170-6.el7_9.5 will be installed\n---> Package sgpio.x86_64 0:1.2.0.10-13.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package Arch Version Repository Size\n================================================================================\nInstalling:\n libblockdev-crypto x86_64 2.18-5.el7 rhel 60 k\n libblockdev-dm x86_64 2.18-5.el7 rhel-optional 54 k\n libblockdev-lvm x86_64 2.18-5.el7 rhel 62 k\n libblockdev-mdraid x86_64 2.18-5.el7 rhel 57 k\n libblockdev-swap x86_64 2.18-5.el7 rhel 52 k\n python-enum34 noarch 1.0.4-1.el7 rhel 52 k\n python2-blivet3 noarch 1:3.1.3-3.el7 rhel 851 k\nInstalling for dependencies:\n blivet3-data noarch 1:3.1.3-3.el7 rhel 77 k\n device-mapper-event\n x86_64 7:1.02.170-6.el7_9.5 rhel 192 k\n device-mapper-event-libs\n x86_64 7:1.02.170-6.el7_9.5 rhel 192 k\n device-mapper-persistent-data\n x86_64 0.8.5-3.el7_9.2 rhel 423 k\n dmraid x86_64 1.0.0.rc16-28.el7 rhel 151 k\n dmraid-events x86_64 1.0.0.rc16-28.el7 rhel 21 k\n json-glib x86_64 1.4.2-2.el7 rhel 134 k\n libaio x86_64 0.3.109-13.el7 rhel 24 k\n libblockdev x86_64 2.18-5.el7 rhel 119 k\n libblockdev-utils x86_64 2.18-5.el7 rhel 59 k\n libbytesize x86_64 1.2-1.el7 rhel 52 k\n libdnf x86_64 0.22.5-2.el7_9 rhel-7-server-extras-rpms 536 k\n libmodulemd x86_64 1.6.3-1.el7 rhel-7-server-extras-rpms 153 k\n librepo x86_64 1.8.1-8.el7_9 rhel 82 k\n libreport-filesystem\n x86_64 2.1.11-53.el7 rhel 41 k\n librhsm x86_64 0.0.3-3.el7_9 rhel-7-server-extras-rpms 28 k\n libsolv x86_64 0.6.34-4.el7 rhel 329 k\n lsof x86_64 4.87-6.el7 rhel 331 k\n lvm2 x86_64 7:2.02.187-6.el7_9.5 rhel 1.3 M\n lvm2-libs x86_64 7:2.02.187-6.el7_9.5 rhel 1.1 M\n mdadm x86_64 4.1-9.el7_9 rhel 440 k\n mpfr x86_64 3.1.1-4.el7 rhel 203 k\n pyparted x86_64 1:3.9-15.el7 rhel 195 k\n python2-blockdev x86_64 2.18-5.el7 rhel 61 k\n python2-bytesize x86_64 1.2-1.el7 rhel 22 k\n python2-hawkey x86_64 0.22.5-2.el7_9 rhel-7-server-extras-rpms 71 k\n python2-libdnf x86_64 0.22.5-2.el7_9 rhel-7-server-extras-rpms 611 k\n sgpio x86_64 1.2.0.10-13.el7 rhel 14 k\n volume_key-libs x86_64 0.3.9-9.el7 rhel 141 k\n\nTransaction Summary\n================================================================================\nInstall 7 Packages (+29 Dependent packages)\n\nTotal download size: 8.2 M\nInstalled size: 24 M\nDownloading packages:\n--------------------------------------------------------------------------------\nTotal 19 MB/s | 8.2 MB 00:00 \nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n Installing : libblockdev-utils-2.18-5.el7.x86_64 1/36 \n Installing : 7:device-mapper-event-libs-1.02.170-6.el7_9.5.x86_64 2/36 \n Installing : json-glib-1.4.2-2.el7.x86_64 3/36 \n Installing : librhsm-0.0.3-3.el7_9.x86_64 4/36 \n Installing : libsolv-0.6.34-4.el7.x86_64 5/36 \n Installing : libaio-0.3.109-13.el7.x86_64 6/36 \n Installing : librepo-1.8.1-8.el7_9.x86_64 7/36 \n Installing : libmodulemd-1.6.3-1.el7.x86_64 8/36 \n Installing : libdnf-0.22.5-2.el7_9.x86_64 9/36 \n Installing : device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64 10/36 \n Installing : 7:device-mapper-event-1.02.170-6.el7_9.5.x86_64 11/36 \n Installing : 7:lvm2-libs-2.02.187-6.el7_9.5.x86_64 12/36 \n Installing : 7:lvm2-2.02.187-6.el7_9.5.x86_64 13/36 \n Installing : python2-libdnf-0.22.5-2.el7_9.x86_64 14/36 \n Installing : python2-hawkey-0.22.5-2.el7_9.x86_64 15/36 \n Installing : libblockdev-2.18-5.el7.x86_64 16/36 \n Installing : python2-blockdev-2.18-5.el7.x86_64 17/36 \n Installing : 1:pyparted-3.9-15.el7.x86_64 18/36 \n Installing : sgpio-1.2.0.10-13.el7.x86_64 19/36 \n Installing : dmraid-1.0.0.rc16-28.el7.x86_64 20/36 \n Installing : dmraid-events-1.0.0.rc16-28.el7.x86_64 21/36 \n Installing : volume_key-libs-0.3.9-9.el7.x86_64 22/36 \n Installing : mpfr-3.1.1-4.el7.x86_64 23/36 \n Installing : libbytesize-1.2-1.el7.x86_64 24/36 \n Installing : python2-bytesize-1.2-1.el7.x86_64 25/36 \n Installing : libreport-filesystem-2.1.11-53.el7.x86_64 26/36 \n Installing : mdadm-4.1-9.el7_9.x86_64 27/36 \n Installing : 1:blivet3-data-3.1.3-3.el7.noarch 28/36 \n Installing : lsof-4.87-6.el7.x86_64 29/36 \n Installing : 1:python2-blivet3-3.1.3-3.el7.noarch 30/36 \n Installing : libblockdev-mdraid-2.18-5.el7.x86_64 31/36 \n Installing : libblockdev-crypto-2.18-5.el7.x86_64 32/36 \n Installing : libblockdev-dm-2.18-5.el7.x86_64 33/36 \n Installing : libblockdev-lvm-2.18-5.el7.x86_64 34/36 \n Installing : libblockdev-swap-2.18-5.el7.x86_64 35/36 \n Installing : python-enum34-1.0.4-1.el7.noarch 36/36 \n Verifying : 7:device-mapper-event-1.02.170-6.el7_9.5.x86_64 1/36 \n Verifying : libblockdev-swap-2.18-5.el7.x86_64 2/36 \n Verifying : librhsm-0.0.3-3.el7_9.x86_64 3/36 \n Verifying : libblockdev-lvm-2.18-5.el7.x86_64 4/36 \n Verifying : lsof-4.87-6.el7.x86_64 5/36 \n Verifying : libblockdev-mdraid-2.18-5.el7.x86_64 6/36 \n Verifying : libdnf-0.22.5-2.el7_9.x86_64 7/36 \n Verifying : python-enum34-1.0.4-1.el7.noarch 8/36 \n Verifying : 1:blivet3-data-3.1.3-3.el7.noarch 9/36 \n Verifying : dmraid-events-1.0.0.rc16-28.el7.x86_64 10/36 \n Verifying : python2-blockdev-2.18-5.el7.x86_64 11/36 \n Verifying : libmodulemd-1.6.3-1.el7.x86_64 12/36 \n Verifying : librepo-1.8.1-8.el7_9.x86_64 13/36 \n Verifying : libblockdev-dm-2.18-5.el7.x86_64 14/36 \n Verifying : json-glib-1.4.2-2.el7.x86_64 15/36 \n Verifying : libaio-0.3.109-13.el7.x86_64 16/36 \n Verifying : 7:lvm2-libs-2.02.187-6.el7_9.5.x86_64 17/36 \n Verifying : python2-hawkey-0.22.5-2.el7_9.x86_64 18/36 \n Verifying : python2-bytesize-1.2-1.el7.x86_64 19/36 \n Verifying : libblockdev-2.18-5.el7.x86_64 20/36 \n Verifying : libreport-filesystem-2.1.11-53.el7.x86_64 21/36 \n Verifying : libbytesize-1.2-1.el7.x86_64 22/36 \n Verifying : 7:device-mapper-event-libs-1.02.170-6.el7_9.5.x86_64 23/36 \n Verifying : python2-libdnf-0.22.5-2.el7_9.x86_64 24/36 \n Verifying : 7:lvm2-2.02.187-6.el7_9.5.x86_64 25/36 \n Verifying : libblockdev-utils-2.18-5.el7.x86_64 26/36 \n Verifying : mpfr-3.1.1-4.el7.x86_64 27/36 \n Verifying : volume_key-libs-0.3.9-9.el7.x86_64 28/36 \n Verifying : libsolv-0.6.34-4.el7.x86_64 29/36 \n Verifying : device-mapper-persistent-data-0.8.5-3.el7_9.2.x86_64 30/36 \n Verifying : 1:python2-blivet3-3.1.3-3.el7.noarch 31/36 \n Verifying : dmraid-1.0.0.rc16-28.el7.x86_64 32/36 \n Verifying : mdadm-4.1-9.el7_9.x86_64 33/36 \n Verifying : sgpio-1.2.0.10-13.el7.x86_64 34/36 \n Verifying : libblockdev-crypto-2.18-5.el7.x86_64 35/36 \n Verifying : 1:pyparted-3.9-15.el7.x86_64 36/36 \n\nInstalled:\n libblockdev-crypto.x86_64 0:2.18-5.el7 libblockdev-dm.x86_64 0:2.18-5.el7 \n libblockdev-lvm.x86_64 0:2.18-5.el7 libblockdev-mdraid.x86_64 0:2.18-5.el7\n libblockdev-swap.x86_64 0:2.18-5.el7 python-enum34.noarch 0:1.0.4-1.el7 \n python2-blivet3.noarch 1:3.1.3-3.el7 \n\nDependency Installed:\n blivet3-data.noarch 1:3.1.3-3.el7 \n device-mapper-event.x86_64 7:1.02.170-6.el7_9.5 \n device-mapper-event-libs.x86_64 7:1.02.170-6.el7_9.5 \n device-mapper-persistent-data.x86_64 0:0.8.5-3.el7_9.2 \n dmraid.x86_64 0:1.0.0.rc16-28.el7 \n dmraid-events.x86_64 0:1.0.0.rc16-28.el7 \n json-glib.x86_64 0:1.4.2-2.el7 \n libaio.x86_64 0:0.3.109-13.el7 \n libblockdev.x86_64 0:2.18-5.el7 \n libblockdev-utils.x86_64 0:2.18-5.el7 \n libbytesize.x86_64 0:1.2-1.el7 \n libdnf.x86_64 0:0.22.5-2.el7_9 \n libmodulemd.x86_64 0:1.6.3-1.el7 \n librepo.x86_64 0:1.8.1-8.el7_9 \n libreport-filesystem.x86_64 0:2.1.11-53.el7 \n librhsm.x86_64 0:0.0.3-3.el7_9 \n libsolv.x86_64 0:0.6.34-4.el7 \n lsof.x86_64 0:4.87-6.el7 \n lvm2.x86_64 7:2.02.187-6.el7_9.5 \n lvm2-libs.x86_64 7:2.02.187-6.el7_9.5 \n mdadm.x86_64 0:4.1-9.el7_9 \n mpfr.x86_64 0:3.1.1-4.el7 \n pyparted.x86_64 1:3.9-15.el7 \n python2-blockdev.x86_64 0:2.18-5.el7 \n python2-bytesize.x86_64 0:1.2-1.el7 \n python2-hawkey.x86_64 0:0.22.5-2.el7_9 \n python2-libdnf.x86_64 0:0.22.5-2.el7_9 \n sgpio.x86_64 0:1.2.0.10-13.el7 \n volume_key-libs.x86_64 0:0.3.9-9.el7 \n\nComplete!\n" ] } TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 18:33:29 +0000 (0:00:09.368) 0:00:13.290 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 18:33:29 +0000 (0:00:00.066) 0:00:13.357 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 18:33:30 +0000 (0:00:00.035) 0:00:13.392 ********* ok: [/cache/rhel-7.qcow2] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 18:33:30 +0000 (0:00:00.630) 0:00:14.022 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 18:33:30 +0000 (0:00:00.046) 0:00:14.068 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 18:33:30 +0000 (0:00:00.033) 0:00:14.102 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 18:33:30 +0000 (0:00:00.039) 0:00:14.141 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 18:33:30 +0000 (0:00:00.075) 0:00:14.217 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [] } TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 18:33:31 +0000 (0:00:00.543) 0:00:14.760 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "brandbot.service": { "name": "brandbot.service", "source": "systemd", "state": "stopped", "status": "static" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-shell.service": { "name": "console-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.import1.service": { "name": "dbus-org.freedesktop.import1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.machine1.service": { "name": "dbus-org.freedesktop.machine1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dmraid-activation.service": { "name": "dmraid-activation.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "exim.service": { "name": "exim.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "ip6tables.service": { "name": "ip6tables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iptables.service": { "name": "iptables.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-activation.service": { "name": "lvm2-activation.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmetad.service": { "name": "lvm2-lvmetad.service", "source": "systemd", "state": "running", "status": "static" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "netconsole": { "name": "netconsole", "source": "sysv", "state": "stopped", "status": "disabled" }, "network": { "name": "network", "source": "sysv", "state": "running", "status": "enabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "active" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-config.service": { "name": "nfs-config.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-idmap.service": { "name": "nfs-idmap.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-lock.service": { "name": "nfs-lock.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-rquotad.service": { "name": "nfs-rquotad.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-secure-server.service": { "name": "nfs-secure-server.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-secure.service": { "name": "nfs-secure.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs.service": { "name": "nfs.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfslock.service": { "name": "nfslock.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ovirt-guest-agent.service": { "name": "ovirt-guest-agent.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "postfix.service": { "name": "postfix.service", "source": "systemd", "state": "running", "status": "enabled" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhcd.service": { "name": "rhcd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhel-autorelabel-mark.service": { "name": "rhel-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-autorelabel.service": { "name": "rhel-autorelabel.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-configure.service": { "name": "rhel-configure.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-dmesg.service": { "name": "rhel-dmesg.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-domainname.service": { "name": "rhel-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-import-state.service": { "name": "rhel-import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-loadmodules.service": { "name": "rhel-loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-readonly.service": { "name": "rhel-readonly.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhnsd": { "name": "rhnsd", "source": "sysv", "state": "running", "status": "enabled" }, "rhnsd.service": { "name": "rhnsd.service", "source": "systemd", "state": "running", "status": "active" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-rquotad.service": { "name": "rpc-rquotad.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpcgssd.service": { "name": "rpcgssd.service", "source": "systemd", "state": "inactive", "status": "static" }, "rpcidmapd.service": { "name": "rpcidmapd.service", "source": "systemd", "state": "inactive", "status": "static" }, "rsyncd.service": { "name": "rsyncd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyncd@.service": { "name": "rsyncd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-policy-migrate-local-changes@.service": { "name": "selinux-policy-migrate-local-changes@.service", "source": "systemd", "state": "unknown", "status": "static" }, "selinux-policy-migrate-local-changes@targeted.service": { "name": "selinux-policy-migrate-local-changes@targeted.service", "source": "systemd", "state": "stopped", "status": "active" }, "sendmail.service": { "name": "sendmail.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "static" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bootchart.service": { "name": "systemd-bootchart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-importd.service": { "name": "systemd-importd.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machined.service": { "name": "systemd-machined.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-nspawn@.service": { "name": "systemd-nspawn@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-readahead-collect.service": { "name": "systemd-readahead-collect.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-readahead-done.service": { "name": "systemd-readahead-done.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "systemd-readahead-drop.service": { "name": "systemd-readahead-drop.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "systemd-readahead-replay.service": { "name": "systemd-readahead-replay.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill@.service": { "name": "systemd-rfkill@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-shutdownd.service": { "name": "systemd-shutdownd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "wpa_supplicant.service": { "name": "wpa_supplicant.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 18:33:32 +0000 (0:00:01.208) 0:00:15.969 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 18:33:32 +0000 (0:00:00.061) 0:00:16.030 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 18:33:32 +0000 (0:00:00.023) 0:00:16.053 ********* ok: [/cache/rhel-7.qcow2] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 18:33:33 +0000 (0:00:00.475) 0:00:16.529 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 18:33:33 +0000 (0:00:00.036) 0:00:16.566 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 18:33:33 +0000 (0:00:00.022) 0:00:16.588 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 18:33:33 +0000 (0:00:00.035) 0:00:16.624 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 18:33:33 +0000 (0:00:00.036) 0:00:16.660 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 18:33:33 +0000 (0:00:00.035) 0:00:16.695 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 18:33:33 +0000 (0:00:00.038) 0:00:16.733 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 18:33:33 +0000 (0:00:00.023) 0:00:16.757 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 18:33:33 +0000 (0:00:00.033) 0:00:16.791 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 18:33:33 +0000 (0:00:00.023) 0:00:16.814 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658428407.582735, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658201031.524, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 70, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658200515.884, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744071677828413", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 18:33:33 +0000 (0:00:00.445) 0:00:17.259 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 18:33:33 +0000 (0:00:00.022) 0:00:17.281 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [Mark tasks to be skipped] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes.yml:14 Thursday 21 July 2022 18:33:34 +0000 (0:00:00.824) 0:00:18.105 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_skip_checks": [ "blivet_available", "packages_installed", "service_facts" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes.yml:21 Thursday 21 July 2022 18:33:34 +0000 (0:00:00.034) 0:00:18.139 ********* included: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml for /cache/rhel-7.qcow2 TASK [Find unused disks in the system] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml:2 Thursday 21 July 2022 18:33:34 +0000 (0:00:00.035) 0:00:18.175 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "disks": [ "sdb" ] } TASK [Set unused_disks if necessary] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml:9 Thursday 21 July 2022 18:33:35 +0000 (0:00:00.487) 0:00:18.663 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "unused_disks": [ "sdb" ] }, "changed": false } TASK [Exit playbook when there's not enough unused disks in the system] ******** task path: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml:14 Thursday 21 July 2022 18:33:35 +0000 (0:00:00.067) 0:00:18.730 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Print unused disks] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/get_unused_disk.yml:19 Thursday 21 July 2022 18:33:35 +0000 (0:00:00.038) 0:00:18.769 ********* ok: [/cache/rhel-7.qcow2] => { "unused_disks": [ "sdb" ] } TASK [Create three LVM logical volumes under one volume group] ***************** task path: /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes.yml:26 Thursday 21 July 2022 18:33:35 +0000 (0:00:00.089) 0:00:18.858 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 18:33:35 +0000 (0:00:00.042) 0:00:18.901 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 18:33:35 +0000 (0:00:00.033) 0:00:18.935 ********* ok: [/cache/rhel-7.qcow2] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 18:33:35 +0000 (0:00:00.403) 0:00:19.338 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 18:33:36 +0000 (0:00:00.069) 0:00:19.408 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 18:33:36 +0000 (0:00:00.035) 0:00:19.443 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 18:33:36 +0000 (0:00:00.033) 0:00:19.477 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 18:33:36 +0000 (0:00:00.055) 0:00:19.533 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 18:33:36 +0000 (0:00:00.020) 0:00:19.553 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 18:33:36 +0000 (0:00:00.026) 0:00:19.579 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "sdb" ], "name": "foo", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "3g" }, { "mount_point": "/opt/test2", "name": "test2", "size": "3g" }, { "mount_point": "/opt/test3", "name": "test3", "size": "3g" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 18:33:36 +0000 (0:00:00.037) 0:00:19.617 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 18:33:36 +0000 (0:00:00.034) 0:00:19.651 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 18:33:36 +0000 (0:00:00.027) 0:00:19.678 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 18:33:36 +0000 (0:00:00.026) 0:00:19.704 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 18:33:36 +0000 (0:00:00.027) 0:00:19.732 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 18:33:36 +0000 (0:00:00.026) 0:00:19.758 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 18:33:36 +0000 (0:00:00.056) 0:00:19.814 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 18:33:36 +0000 (0:00:00.021) 0:00:19.836 ********* changed: [/cache/rhel-7.qcow2] => { "actions": [ { "action": "create format", "device": "/dev/sdb", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test3", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test3", "fs_type": "xfs" }, { "action": "create device", "device": "/dev/mapper/foo-test2", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test2", "fs_type": "xfs" }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2", "/dev/mapper/foo-test3" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 18:33:38 +0000 (0:00:01.870) 0:00:21.706 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 18:33:38 +0000 (0:00:00.038) 0:00:21.744 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 18:33:38 +0000 (0:00:00.021) 0:00:21.766 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/sdb", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test3", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test3", "fs_type": "xfs" }, { "action": "create device", "device": "/dev/mapper/foo-test2", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test2", "fs_type": "xfs" }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2", "/dev/mapper/foo-test3" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 18:33:38 +0000 (0:00:00.071) 0:00:21.838 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 18:33:38 +0000 (0:00:00.076) 0:00:21.915 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 18:33:38 +0000 (0:00:00.037) 0:00:21.952 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 18:33:38 +0000 (0:00:00.037) 0:00:21.990 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 18:33:39 +0000 (0:00:00.756) 0:00:22.746 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test2', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test2', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test3', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test3', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "mounted" }, "name": "/opt/test3", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test3" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 18:33:40 +0000 (0:00:01.180) 0:00:23.927 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 18:33:40 +0000 (0:00:00.457) 0:00:24.384 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658428407.582735, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658201031.524, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 70, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658200515.884, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744071677828413", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 18:33:41 +0000 (0:00:00.314) 0:00:24.699 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 18:33:41 +0000 (0:00:00.022) 0:00:24.721 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes.yml:44 Thursday 21 July 2022 18:33:42 +0000 (0:00:00.881) 0:00:25.603 ********* included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 18:33:42 +0000 (0:00:00.041) 0:00:25.644 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 18:33:42 +0000 (0:00:00.081) 0:00:25.725 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 18:33:42 +0000 (0:00:00.038) 0:00:25.763 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "3G", "type": "lvm", "uuid": "fc66144f-ab8b-462b-8161-13a4b348250a" }, "/dev/mapper/foo-test2": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test2", "size": "3G", "type": "lvm", "uuid": "fe36895e-c8bf-4620-998e-4265784b74a4" }, "/dev/mapper/foo-test3": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test3", "size": "3G", "type": "lvm", "uuid": "a8832341-3e89-4bdb-8994-bc5d31e2826b" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "fUmW0y-6Mtw-liD0-pqhj-52b5-lflx-H3Il8d" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-18-33-04-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 18:33:43 +0000 (0:00:01.446) 0:00:27.210 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003492", "end": "2022-07-21 14:33:44.221201", "rc": 0, "start": "2022-07-21 14:33:44.217709" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 /dev/mapper/foo-test2 /opt/test2 xfs defaults 0 0 /dev/mapper/foo-test3 /opt/test3 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 18:33:44 +0000 (0:00:00.468) 0:00:27.678 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003463", "end": "2022-07-21 14:33:44.579370", "failed_when_result": false, "rc": 0, "start": "2022-07-21 14:33:44.575907" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 18:33:44 +0000 (0:00:00.358) 0:00:28.037 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, {'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, {'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 18:33:44 +0000 (0:00:00.099) 0:00:28.136 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 18:33:44 +0000 (0:00:00.098) 0:00:28.235 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members) included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 18:33:44 +0000 (0:00:00.083) 0:00:28.319 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sdb" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.070) 0:00:28.390 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb", "pv": "/dev/sdb" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.414) 0:00:28.804 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.049) 0:00:28.853 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sdb" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.050) 0:00:28.903 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.049) 0:00:28.953 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.037) 0:00:28.991 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.047) 0:00:29.039 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.025) 0:00:29.064 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.042) 0:00:29.107 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml for /cache/rhel-7.qcow2 TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.044) 0:00:29.151 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.024) 0:00:29.176 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.024) 0:00:29.200 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.022) 0:00:29.223 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.023) 0:00:29.246 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.022) 0:00:29.269 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.022) 0:00:29.291 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.022) 0:00:29.313 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 18:33:45 +0000 (0:00:00.034) 0:00:29.348 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.042) 0:00:29.390 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.053) 0:00:29.443 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.026) 0:00:29.469 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.025) 0:00:29.495 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.026) 0:00:29.522 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.026) 0:00:29.549 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.025) 0:00:29.574 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.025) 0:00:29.600 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.026) 0:00:29.627 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.026) 0:00:29.653 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.026) 0:00:29.680 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml for /cache/rhel-7.qcow2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.072) 0:00:29.752 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.056) 0:00:29.809 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.023) 0:00:29.832 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.023) 0:00:29.856 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.022) 0:00:29.878 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.022) 0:00:29.901 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.022) 0:00:29.923 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.023) 0:00:29.947 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.025) 0:00:29.973 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.026) 0:00:29.999 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.024) 0:00:30.024 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.027) 0:00:30.051 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.027) 0:00:30.078 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.023) 0:00:30.102 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2 TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.050) 0:00:30.153 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.051) 0:00:30.205 ********* skipping: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "_storage_test_pool_member_path": "/dev/sdb", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.028) 0:00:30.234 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/sdb) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.045) 0:00:30.279 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 18:33:46 +0000 (0:00:00.056) 0:00:30.335 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.055) 0:00:30.391 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.039) 0:00:30.431 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.038) 0:00:30.469 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.040) 0:00:30.510 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.039) 0:00:30.549 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.035) 0:00:30.585 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.047) 0:00:30.632 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.060) 0:00:30.693 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.024) 0:00:30.717 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.024) 0:00:30.741 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.023) 0:00:30.765 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.022) 0:00:30.787 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.022) 0:00:30.810 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.024) 0:00:30.835 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.025) 0:00:30.860 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.080) 0:00:30.941 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.069) 0:00:31.010 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.026) 0:00:31.037 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.025) 0:00:31.062 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.023) 0:00:31.085 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.024) 0:00:31.109 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.024) 0:00:31.134 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.022) 0:00:31.157 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.036) 0:00:31.194 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.026) 0:00:31.220 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.024) 0:00:31.244 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.023) 0:00:31.267 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.022) 0:00:31.290 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.024) 0:00:31.314 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.024) 0:00:31.339 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 18:33:47 +0000 (0:00:00.024) 0:00:31.364 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.036) 0:00:31.400 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.037) 0:00:31.438 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.055) 0:00:31.493 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.052) 0:00:31.545 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.084) 0:00:31.630 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.039) 0:00:31.670 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "fc66144f-ab8b-462b-8161-13a4b348250a" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "fc66144f-ab8b-462b-8161-13a4b348250a" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.056) 0:00:31.726 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.050) 0:00:31.777 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.048) 0:00:31.826 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.050) 0:00:31.876 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.026) 0:00:31.903 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.026) 0:00:31.930 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.023) 0:00:31.953 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.033) 0:00:31.987 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.061) 0:00:32.048 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.048) 0:00:32.096 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.048) 0:00:32.144 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.037) 0:00:32.182 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.033) 0:00:32.215 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:33:48 +0000 (0:00:00.093) 0:00:32.309 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:33:49 +0000 (0:00:00.115) 0:00:32.424 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658428418.179735, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658428418.179735, "dev": 5, "device_type": 64514, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 35068, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658428418.179735, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:33:49 +0000 (0:00:00.307) 0:00:32.732 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:33:49 +0000 (0:00:00.036) 0:00:32.768 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:33:49 +0000 (0:00:00.036) 0:00:32.805 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:33:49 +0000 (0:00:00.035) 0:00:32.840 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:33:49 +0000 (0:00:00.022) 0:00:32.863 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:33:49 +0000 (0:00:00.037) 0:00:32.901 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:33:49 +0000 (0:00:00.023) 0:00:32.924 ********* changed: [/cache/rhel-7.qcow2] => { "changed": true, "changes": { "installed": [ "cryptsetup" ] }, "rc": 0, "results": [ "Loaded plugins: search-disabled-repos\nResolving Dependencies\n--> Running transaction check\n---> Package cryptsetup.x86_64 0:2.0.3-6.el7 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package Arch Version Repository Size\n================================================================================\nInstalling:\n cryptsetup x86_64 2.0.3-6.el7 rhel 154 k\n\nTransaction Summary\n================================================================================\nInstall 1 Package\n\nTotal download size: 154 k\nInstalled size: 354 k\nDownloading packages:\nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n Installing : cryptsetup-2.0.3-6.el7.x86_64 1/1 \n Verifying : cryptsetup-2.0.3-6.el7.x86_64 1/1 \n\nInstalled:\n cryptsetup.x86_64 0:2.0.3-6.el7 \n\nComplete!\n" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:33:51 +0000 (0:00:01.546) 0:00:34.471 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.024) 0:00:34.495 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.023) 0:00:34.518 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.051) 0:00:34.570 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.025) 0:00:34.595 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.023) 0:00:34.619 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.024) 0:00:34.643 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.023) 0:00:34.667 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.024) 0:00:34.692 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.048) 0:00:34.740 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.048) 0:00:34.789 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.037) 0:00:34.826 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.034) 0:00:34.861 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.034) 0:00:34.895 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.033) 0:00:34.929 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.036) 0:00:34.965 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.035) 0:00:35.000 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.035) 0:00:35.036 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.034) 0:00:35.071 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.037) 0:00:35.108 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.035) 0:00:35.144 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:33:51 +0000 (0:00:00.036) 0:00:35.180 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:33:52 +0000 (0:00:00.425) 0:00:35.605 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:33:52 +0000 (0:00:00.376) 0:00:35.982 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:33:52 +0000 (0:00:00.064) 0:00:36.046 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:33:52 +0000 (0:00:00.038) 0:00:36.085 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:33:52 +0000 (0:00:00.048) 0:00:36.133 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:33:52 +0000 (0:00:00.039) 0:00:36.172 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:33:52 +0000 (0:00:00.039) 0:00:36.212 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:33:52 +0000 (0:00:00.038) 0:00:36.251 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:33:52 +0000 (0:00:00.039) 0:00:36.291 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:33:52 +0000 (0:00:00.035) 0:00:36.326 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:33:52 +0000 (0:00:00.037) 0:00:36.364 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:33:53 +0000 (0:00:00.054) 0:00:36.419 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.036677", "end": "2022-07-21 14:33:53.303191", "rc": 0, "start": "2022-07-21 14:33:53.266514" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:33:53 +0000 (0:00:00.351) 0:00:36.770 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:33:53 +0000 (0:00:00.048) 0:00:36.818 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:33:53 +0000 (0:00:00.049) 0:00:36.868 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:33:53 +0000 (0:00:00.038) 0:00:36.906 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:33:53 +0000 (0:00:00.035) 0:00:36.942 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:33:53 +0000 (0:00:00.034) 0:00:36.977 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:33:53 +0000 (0:00:00.034) 0:00:37.011 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:33:53 +0000 (0:00:00.034) 0:00:37.046 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:33:53 +0000 (0:00:00.046) 0:00:37.093 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:33:53 +0000 (0:00:00.082) 0:00:37.175 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:33:53 +0000 (0:00:00.046) 0:00:37.221 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test2", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "fe36895e-c8bf-4620-998e-4265784b74a4" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test2", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "fe36895e-c8bf-4620-998e-4265784b74a4" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:33:53 +0000 (0:00:00.055) 0:00:37.277 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:33:53 +0000 (0:00:00.090) 0:00:37.367 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:33:54 +0000 (0:00:00.049) 0:00:37.417 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:33:54 +0000 (0:00:00.049) 0:00:37.466 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:33:54 +0000 (0:00:00.023) 0:00:37.490 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:33:54 +0000 (0:00:00.023) 0:00:37.514 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:33:54 +0000 (0:00:00.024) 0:00:37.538 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:33:54 +0000 (0:00:00.032) 0:00:37.571 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test2 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:33:54 +0000 (0:00:00.096) 0:00:37.668 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:33:54 +0000 (0:00:00.118) 0:00:37.786 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:33:54 +0000 (0:00:00.048) 0:00:37.834 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:33:54 +0000 (0:00:00.035) 0:00:37.869 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:33:54 +0000 (0:00:00.034) 0:00:37.904 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:33:54 +0000 (0:00:00.036) 0:00:37.941 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:33:54 +0000 (0:00:00.038) 0:00:37.979 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658428418.003735, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658428418.003735, "dev": 5, "device_type": 64513, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 35038, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658428418.003735, "nlink": 1, "path": "/dev/mapper/foo-test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:33:54 +0000 (0:00:00.367) 0:00:38.347 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:33:54 +0000 (0:00:00.036) 0:00:38.384 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:33:55 +0000 (0:00:00.036) 0:00:38.420 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:33:55 +0000 (0:00:00.035) 0:00:38.455 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:33:55 +0000 (0:00:00.023) 0:00:38.478 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:33:55 +0000 (0:00:00.039) 0:00:38.518 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:33:55 +0000 (0:00:00.023) 0:00:38.541 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:33:55 +0000 (0:00:00.554) 0:00:39.096 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:33:55 +0000 (0:00:00.025) 0:00:39.121 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:33:55 +0000 (0:00:00.023) 0:00:39.145 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:33:55 +0000 (0:00:00.054) 0:00:39.199 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:33:55 +0000 (0:00:00.022) 0:00:39.222 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:33:55 +0000 (0:00:00.022) 0:00:39.245 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:33:55 +0000 (0:00:00.022) 0:00:39.267 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:33:55 +0000 (0:00:00.024) 0:00:39.292 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:33:55 +0000 (0:00:00.023) 0:00:39.315 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:33:55 +0000 (0:00:00.048) 0:00:39.364 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:33:56 +0000 (0:00:00.050) 0:00:39.414 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:33:56 +0000 (0:00:00.038) 0:00:39.452 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:33:56 +0000 (0:00:00.038) 0:00:39.490 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:33:56 +0000 (0:00:00.038) 0:00:39.529 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:33:56 +0000 (0:00:00.039) 0:00:39.568 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:33:56 +0000 (0:00:00.038) 0:00:39.606 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:33:56 +0000 (0:00:00.037) 0:00:39.643 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:33:56 +0000 (0:00:00.034) 0:00:39.678 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:33:56 +0000 (0:00:00.035) 0:00:39.714 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:33:56 +0000 (0:00:00.068) 0:00:39.782 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:33:56 +0000 (0:00:00.036) 0:00:39.819 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:33:56 +0000 (0:00:00.036) 0:00:39.855 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:33:56 +0000 (0:00:00.301) 0:00:40.156 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:33:57 +0000 (0:00:00.318) 0:00:40.475 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:33:57 +0000 (0:00:00.051) 0:00:40.526 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:33:57 +0000 (0:00:00.035) 0:00:40.562 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:33:57 +0000 (0:00:00.036) 0:00:40.599 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:33:57 +0000 (0:00:00.040) 0:00:40.640 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:33:57 +0000 (0:00:00.036) 0:00:40.676 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:33:57 +0000 (0:00:00.036) 0:00:40.713 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:33:57 +0000 (0:00:00.036) 0:00:40.749 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:33:57 +0000 (0:00:00.040) 0:00:40.790 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:33:57 +0000 (0:00:00.033) 0:00:40.824 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:33:57 +0000 (0:00:00.050) 0:00:40.874 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test2" ], "delta": "0:00:00.024537", "end": "2022-07-21 14:33:57.738559", "rc": 0, "start": "2022-07-21 14:33:57.714022" } STDOUT: LVM2_LV_NAME=test2 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:33:57 +0000 (0:00:00.320) 0:00:41.194 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:33:57 +0000 (0:00:00.047) 0:00:41.242 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:33:57 +0000 (0:00:00.051) 0:00:41.293 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:33:57 +0000 (0:00:00.038) 0:00:41.332 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:33:57 +0000 (0:00:00.039) 0:00:41.372 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.036) 0:00:41.409 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.036) 0:00:41.445 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.036) 0:00:41.481 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.051) 0:00:41.533 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.078) 0:00:41.611 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test3" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.039) 0:00:41.651 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test3", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test3", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "a8832341-3e89-4bdb-8994-bc5d31e2826b" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test3", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test3", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "a8832341-3e89-4bdb-8994-bc5d31e2826b" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.097) 0:00:41.748 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.050) 0:00:41.799 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.083) 0:00:41.882 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.085) 0:00:41.967 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.023) 0:00:41.990 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.022) 0:00:42.013 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.022) 0:00:42.036 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.093) 0:00:42.129 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test3 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test3 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test3 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.059) 0:00:42.189 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.051) 0:00:42.240 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.050) 0:00:42.291 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.038) 0:00:42.329 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:33:58 +0000 (0:00:00.036) 0:00:42.366 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:33:59 +0000 (0:00:00.041) 0:00:42.408 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:33:59 +0000 (0:00:00.041) 0:00:42.449 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658428417.825735, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658428417.825735, "dev": 5, "device_type": 64512, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 34991, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658428417.825735, "nlink": 1, "path": "/dev/mapper/foo-test3", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:33:59 +0000 (0:00:00.333) 0:00:42.783 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:33:59 +0000 (0:00:00.038) 0:00:42.821 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:33:59 +0000 (0:00:00.039) 0:00:42.861 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:33:59 +0000 (0:00:00.035) 0:00:42.896 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:33:59 +0000 (0:00:00.025) 0:00:42.921 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:33:59 +0000 (0:00:00.039) 0:00:42.961 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:33:59 +0000 (0:00:00.026) 0:00:42.988 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.538) 0:00:43.527 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.027) 0:00:43.554 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.026) 0:00:43.581 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.058) 0:00:43.639 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.026) 0:00:43.666 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.024) 0:00:43.690 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.024) 0:00:43.715 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.034) 0:00:43.750 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.031) 0:00:43.781 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.052) 0:00:43.834 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.048) 0:00:43.883 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.034) 0:00:43.918 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.038) 0:00:43.957 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.039) 0:00:43.996 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.073) 0:00:44.069 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.080) 0:00:44.150 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.037) 0:00:44.187 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.036) 0:00:44.224 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.038) 0:00:44.262 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.039) 0:00:44.302 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.038) 0:00:44.341 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:34:00 +0000 (0:00:00.039) 0:00:44.380 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:34:01 +0000 (0:00:00.323) 0:00:44.704 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:34:01 +0000 (0:00:00.342) 0:00:45.046 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:34:01 +0000 (0:00:00.051) 0:00:45.098 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:34:01 +0000 (0:00:00.036) 0:00:45.134 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:34:01 +0000 (0:00:00.038) 0:00:45.173 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:34:01 +0000 (0:00:00.039) 0:00:45.212 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:34:01 +0000 (0:00:00.042) 0:00:45.255 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:34:01 +0000 (0:00:00.039) 0:00:45.294 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:34:01 +0000 (0:00:00.038) 0:00:45.332 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:34:01 +0000 (0:00:00.039) 0:00:45.372 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:34:02 +0000 (0:00:00.035) 0:00:45.407 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:34:02 +0000 (0:00:00.053) 0:00:45.460 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test3" ], "delta": "0:00:00.030355", "end": "2022-07-21 14:34:02.391059", "rc": 0, "start": "2022-07-21 14:34:02.360704" } STDOUT: LVM2_LV_NAME=test3 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:34:02 +0000 (0:00:00.393) 0:00:45.854 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:34:02 +0000 (0:00:00.050) 0:00:45.905 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:34:02 +0000 (0:00:00.053) 0:00:45.958 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:34:02 +0000 (0:00:00.038) 0:00:45.996 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:34:02 +0000 (0:00:00.040) 0:00:46.036 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:34:02 +0000 (0:00:00.037) 0:00:46.073 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:34:02 +0000 (0:00:00.037) 0:00:46.111 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 18:34:02 +0000 (0:00:00.035) 0:00:46.147 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 18:34:02 +0000 (0:00:00.022) 0:00:46.170 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Repeat the previous invocation to verify idempotence] ******************** task path: /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes.yml:46 Thursday 21 July 2022 18:34:02 +0000 (0:00:00.033) 0:00:46.203 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 18:34:02 +0000 (0:00:00.075) 0:00:46.278 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 18:34:02 +0000 (0:00:00.037) 0:00:46.315 ********* ok: [/cache/rhel-7.qcow2] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 18:34:03 +0000 (0:00:00.416) 0:00:46.731 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 18:34:03 +0000 (0:00:00.059) 0:00:46.791 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 18:34:03 +0000 (0:00:00.034) 0:00:46.825 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 18:34:03 +0000 (0:00:00.033) 0:00:46.859 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 18:34:03 +0000 (0:00:00.055) 0:00:46.914 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 18:34:03 +0000 (0:00:00.019) 0:00:46.934 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 18:34:03 +0000 (0:00:00.026) 0:00:46.961 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "sdb" ], "name": "foo", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "3g" }, { "mount_point": "/opt/test2", "name": "test2", "size": "3g" }, { "mount_point": "/opt/test3", "name": "test3", "size": "3g" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 18:34:03 +0000 (0:00:00.034) 0:00:46.996 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 18:34:03 +0000 (0:00:00.036) 0:00:47.033 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 18:34:03 +0000 (0:00:00.027) 0:00:47.060 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 18:34:03 +0000 (0:00:00.028) 0:00:47.089 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 18:34:03 +0000 (0:00:00.027) 0:00:47.116 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 18:34:03 +0000 (0:00:00.027) 0:00:47.143 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 18:34:03 +0000 (0:00:00.056) 0:00:47.200 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 18:34:03 +0000 (0:00:00.022) 0:00:47.223 ********* ok: [/cache/rhel-7.qcow2] => { "actions": [], "changed": false, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/mapper/foo-test2", "/dev/mapper/foo-test3", "/dev/mapper/foo-test1", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 18:34:05 +0000 (0:00:01.723) 0:00:48.946 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 18:34:05 +0000 (0:00:00.035) 0:00:48.982 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 18:34:05 +0000 (0:00:00.020) 0:00:49.003 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/mapper/foo-test2", "/dev/mapper/foo-test3", "/dev/mapper/foo-test1", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 18:34:05 +0000 (0:00:00.071) 0:00:49.074 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 18:34:05 +0000 (0:00:00.043) 0:00:49.117 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 18:34:05 +0000 (0:00:00.039) 0:00:49.156 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 18:34:05 +0000 (0:00:00.038) 0:00:49.195 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 18:34:06 +0000 (0:00:00.537) 0:00:49.732 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test2', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test2', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test3', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test3', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "mounted" }, "name": "/opt/test3", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test3" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 18:34:07 +0000 (0:00:00.940) 0:00:50.672 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 18:34:07 +0000 (0:00:00.472) 0:00:51.145 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658428407.582735, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658201031.524, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 70, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658200515.884, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744071677828413", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 18:34:08 +0000 (0:00:00.315) 0:00:51.461 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 18:34:08 +0000 (0:00:00.023) 0:00:51.485 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes.yml:64 Thursday 21 July 2022 18:34:09 +0000 (0:00:01.859) 0:00:53.344 ********* included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 18:34:09 +0000 (0:00:00.042) 0:00:53.387 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "_raw_kernel_device": "/dev/dm-1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 18:34:10 +0000 (0:00:00.051) 0:00:53.438 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 18:34:10 +0000 (0:00:00.035) 0:00:53.474 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "3G", "type": "lvm", "uuid": "fc66144f-ab8b-462b-8161-13a4b348250a" }, "/dev/mapper/foo-test2": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test2", "size": "3G", "type": "lvm", "uuid": "fe36895e-c8bf-4620-998e-4265784b74a4" }, "/dev/mapper/foo-test3": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test3", "size": "3G", "type": "lvm", "uuid": "a8832341-3e89-4bdb-8994-bc5d31e2826b" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "fUmW0y-6Mtw-liD0-pqhj-52b5-lflx-H3Il8d" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-18-33-04-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 18:34:10 +0000 (0:00:00.315) 0:00:53.789 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003644", "end": "2022-07-21 14:34:10.631049", "rc": 0, "start": "2022-07-21 14:34:10.627405" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 /dev/mapper/foo-test2 /opt/test2 xfs defaults 0 0 /dev/mapper/foo-test3 /opt/test3 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 18:34:10 +0000 (0:00:00.297) 0:00:54.087 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003486", "end": "2022-07-21 14:34:10.937740", "failed_when_result": false, "rc": 0, "start": "2022-07-21 14:34:10.934254" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 18:34:11 +0000 (0:00:00.308) 0:00:54.395 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, {'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, {'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 18:34:11 +0000 (0:00:00.058) 0:00:54.454 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 18:34:11 +0000 (0:00:00.062) 0:00:54.517 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members) included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 18:34:11 +0000 (0:00:00.045) 0:00:54.563 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sdb" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 18:34:11 +0000 (0:00:00.081) 0:00:54.644 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb", "pv": "/dev/sdb" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 18:34:11 +0000 (0:00:00.340) 0:00:54.985 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 18:34:11 +0000 (0:00:00.078) 0:00:55.063 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sdb" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 18:34:11 +0000 (0:00:00.078) 0:00:55.142 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 18:34:11 +0000 (0:00:00.093) 0:00:55.235 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 18:34:11 +0000 (0:00:00.036) 0:00:55.271 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 18:34:11 +0000 (0:00:00.048) 0:00:55.319 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 18:34:11 +0000 (0:00:00.023) 0:00:55.343 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 18:34:11 +0000 (0:00:00.040) 0:00:55.384 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml for /cache/rhel-7.qcow2 TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.040) 0:00:55.424 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.023) 0:00:55.448 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.022) 0:00:55.471 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.022) 0:00:55.494 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.022) 0:00:55.517 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.023) 0:00:55.540 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.022) 0:00:55.563 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.022) 0:00:55.586 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.036) 0:00:55.622 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.045) 0:00:55.668 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.056) 0:00:55.724 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.026) 0:00:55.751 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.026) 0:00:55.777 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.026) 0:00:55.803 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.027) 0:00:55.831 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.026) 0:00:55.857 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.025) 0:00:55.883 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.026) 0:00:55.909 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.027) 0:00:55.936 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.027) 0:00:55.963 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml for /cache/rhel-7.qcow2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.043) 0:00:56.007 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.058) 0:00:56.065 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.024) 0:00:56.089 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.023) 0:00:56.113 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.022) 0:00:56.135 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.022) 0:00:56.158 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.022) 0:00:56.180 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.023) 0:00:56.204 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.023) 0:00:56.228 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.022) 0:00:56.250 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.022) 0:00:56.273 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.051) 0:00:56.324 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.024) 0:00:56.348 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 18:34:12 +0000 (0:00:00.024) 0:00:56.373 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2 TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.049) 0:00:56.422 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.051) 0:00:56.474 ********* skipping: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "_storage_test_pool_member_path": "/dev/sdb", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.029) 0:00:56.504 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/sdb) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.046) 0:00:56.551 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.049) 0:00:56.600 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.051) 0:00:56.652 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.040) 0:00:56.693 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.036) 0:00:56.729 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.037) 0:00:56.766 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.042) 0:00:56.808 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.037) 0:00:56.846 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.047) 0:00:56.894 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.060) 0:00:56.954 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.024) 0:00:56.979 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.023) 0:00:57.002 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.024) 0:00:57.027 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.023) 0:00:57.050 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.023) 0:00:57.074 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.022) 0:00:57.096 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.026) 0:00:57.123 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.035) 0:00:57.159 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.024) 0:00:57.183 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.024) 0:00:57.208 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.025) 0:00:57.233 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.023) 0:00:57.256 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.023) 0:00:57.279 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.023) 0:00:57.302 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.025) 0:00:57.328 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.032) 0:00:57.361 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 18:34:13 +0000 (0:00:00.024) 0:00:57.385 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.023) 0:00:57.408 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.025) 0:00:57.433 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.023) 0:00:57.457 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.023) 0:00:57.480 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.023) 0:00:57.503 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.059) 0:00:57.563 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.032) 0:00:57.595 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.033) 0:00:57.628 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-1', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-1', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-0', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-0', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.051) 0:00:57.680 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.047) 0:00:57.728 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.077) 0:00:57.805 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.039) 0:00:57.845 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "fc66144f-ab8b-462b-8161-13a4b348250a" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "fc66144f-ab8b-462b-8161-13a4b348250a" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.054) 0:00:57.899 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.048) 0:00:57.948 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.047) 0:00:57.995 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.048) 0:00:58.043 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.023) 0:00:58.067 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.023) 0:00:58.090 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.024) 0:00:58.114 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.032) 0:00:58.146 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.059) 0:00:58.206 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.049) 0:00:58.255 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.048) 0:00:58.304 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.035) 0:00:58.339 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:34:14 +0000 (0:00:00.034) 0:00:58.374 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:34:15 +0000 (0:00:00.039) 0:00:58.413 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:34:15 +0000 (0:00:00.038) 0:00:58.451 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658428418.179735, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658428418.179735, "dev": 5, "device_type": 64514, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 35068, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658428418.179735, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:34:15 +0000 (0:00:00.311) 0:00:58.763 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:34:15 +0000 (0:00:00.036) 0:00:58.800 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:34:15 +0000 (0:00:00.037) 0:00:58.837 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:34:15 +0000 (0:00:00.035) 0:00:58.873 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:34:15 +0000 (0:00:00.022) 0:00:58.895 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:34:15 +0000 (0:00:00.037) 0:00:58.932 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:34:15 +0000 (0:00:00.023) 0:00:58.955 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.519) 0:00:59.475 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.024) 0:00:59.500 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.052) 0:00:59.552 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.049) 0:00:59.602 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.022) 0:00:59.625 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.022) 0:00:59.648 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.022) 0:00:59.670 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.022) 0:00:59.693 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.023) 0:00:59.716 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.050) 0:00:59.767 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.050) 0:00:59.818 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.035) 0:00:59.853 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.035) 0:00:59.889 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.038) 0:00:59.927 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.033) 0:00:59.960 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.035) 0:00:59.995 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.035) 0:01:00.031 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.038) 0:01:00.070 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.035) 0:01:00.106 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.035) 0:01:00.141 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.035) 0:01:00.177 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:34:16 +0000 (0:00:00.040) 0:01:00.218 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:34:17 +0000 (0:00:00.293) 0:01:00.511 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:34:17 +0000 (0:00:00.316) 0:01:00.827 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:34:17 +0000 (0:00:00.049) 0:01:00.876 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:34:17 +0000 (0:00:00.035) 0:01:00.912 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:34:17 +0000 (0:00:00.034) 0:01:00.947 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:34:17 +0000 (0:00:00.035) 0:01:00.982 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:34:17 +0000 (0:00:00.034) 0:01:01.017 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:34:17 +0000 (0:00:00.035) 0:01:01.053 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:34:17 +0000 (0:00:00.034) 0:01:01.088 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:34:17 +0000 (0:00:00.034) 0:01:01.123 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:34:17 +0000 (0:00:00.036) 0:01:01.159 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:34:17 +0000 (0:00:00.052) 0:01:01.211 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.022879", "end": "2022-07-21 14:34:18.076770", "rc": 0, "start": "2022-07-21 14:34:18.053891" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.322) 0:01:01.534 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.081) 0:01:01.615 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.112) 0:01:01.728 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.041) 0:01:01.770 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.039) 0:01:01.810 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.039) 0:01:01.850 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.041) 0:01:01.891 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.036) 0:01:01.928 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.050) 0:01:01.979 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.079) 0:01:02.058 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.042) 0:01:02.100 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test2", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "fe36895e-c8bf-4620-998e-4265784b74a4" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test2", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "fe36895e-c8bf-4620-998e-4265784b74a4" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.053) 0:01:02.154 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.048) 0:01:02.202 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.046) 0:01:02.249 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.047) 0:01:02.297 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.023) 0:01:02.320 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.023) 0:01:02.343 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:34:18 +0000 (0:00:00.024) 0:01:02.368 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:34:19 +0000 (0:00:00.035) 0:01:02.403 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test2 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:34:19 +0000 (0:00:00.061) 0:01:02.464 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:34:19 +0000 (0:00:00.048) 0:01:02.513 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:34:19 +0000 (0:00:00.047) 0:01:02.560 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:34:19 +0000 (0:00:00.035) 0:01:02.596 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:34:19 +0000 (0:00:00.034) 0:01:02.630 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:34:19 +0000 (0:00:00.040) 0:01:02.670 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:34:19 +0000 (0:00:00.038) 0:01:02.709 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658428418.003735, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658428418.003735, "dev": 5, "device_type": 64513, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 35038, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658428418.003735, "nlink": 1, "path": "/dev/mapper/foo-test2", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:34:19 +0000 (0:00:00.318) 0:01:03.027 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:34:19 +0000 (0:00:00.039) 0:01:03.067 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:34:19 +0000 (0:00:00.038) 0:01:03.105 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:34:19 +0000 (0:00:00.038) 0:01:03.144 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:34:19 +0000 (0:00:00.024) 0:01:03.168 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:34:19 +0000 (0:00:00.042) 0:01:03.210 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:34:19 +0000 (0:00:00.026) 0:01:03.236 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.513) 0:01:03.750 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.026) 0:01:03.776 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.053) 0:01:03.830 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.050) 0:01:03.880 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.023) 0:01:03.904 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.024) 0:01:03.929 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.023) 0:01:03.953 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.025) 0:01:03.979 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.025) 0:01:04.005 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.055) 0:01:04.061 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.051) 0:01:04.112 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.049) 0:01:04.162 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.038) 0:01:04.200 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.036) 0:01:04.236 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.034) 0:01:04.271 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.036) 0:01:04.308 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.038) 0:01:04.346 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:34:20 +0000 (0:00:00.036) 0:01:04.383 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:34:21 +0000 (0:00:00.036) 0:01:04.419 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:34:21 +0000 (0:00:00.036) 0:01:04.456 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:34:21 +0000 (0:00:00.037) 0:01:04.494 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:34:21 +0000 (0:00:00.036) 0:01:04.530 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:34:21 +0000 (0:00:00.309) 0:01:04.839 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:34:21 +0000 (0:00:00.312) 0:01:05.152 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:34:21 +0000 (0:00:00.064) 0:01:05.216 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:34:21 +0000 (0:00:00.036) 0:01:05.252 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:34:21 +0000 (0:00:00.034) 0:01:05.287 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:34:21 +0000 (0:00:00.037) 0:01:05.325 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:34:21 +0000 (0:00:00.036) 0:01:05.362 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:34:22 +0000 (0:00:00.037) 0:01:05.399 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:34:22 +0000 (0:00:00.034) 0:01:05.434 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:34:22 +0000 (0:00:00.035) 0:01:05.470 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:34:22 +0000 (0:00:00.037) 0:01:05.507 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:34:22 +0000 (0:00:00.052) 0:01:05.560 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test2" ], "delta": "0:00:00.025453", "end": "2022-07-21 14:34:22.467740", "rc": 0, "start": "2022-07-21 14:34:22.442287" } STDOUT: LVM2_LV_NAME=test2 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:34:22 +0000 (0:00:00.365) 0:01:05.926 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:34:22 +0000 (0:00:00.086) 0:01:06.012 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:34:22 +0000 (0:00:00.080) 0:01:06.093 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:34:22 +0000 (0:00:00.036) 0:01:06.129 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:34:22 +0000 (0:00:00.035) 0:01:06.165 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:34:22 +0000 (0:00:00.036) 0:01:06.201 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:34:22 +0000 (0:00:00.035) 0:01:06.237 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:34:22 +0000 (0:00:00.035) 0:01:06.273 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:34:22 +0000 (0:00:00.051) 0:01:06.324 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.081) 0:01:06.406 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test3" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.044) 0:01:06.450 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test3", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test3", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "a8832341-3e89-4bdb-8994-bc5d31e2826b" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test3", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test3", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "a8832341-3e89-4bdb-8994-bc5d31e2826b" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.061) 0:01:06.511 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.057) 0:01:06.568 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.051) 0:01:06.620 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.050) 0:01:06.671 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.024) 0:01:06.696 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.026) 0:01:06.722 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.024) 0:01:06.747 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.036) 0:01:06.784 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test3 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test3 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test3 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.062) 0:01:06.846 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.052) 0:01:06.899 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.050) 0:01:06.949 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.036) 0:01:06.986 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.038) 0:01:07.024 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.037) 0:01:07.062 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:34:23 +0000 (0:00:00.040) 0:01:07.102 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658428417.825735, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658428417.825735, "dev": 5, "device_type": 64512, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 34991, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658428417.825735, "nlink": 1, "path": "/dev/mapper/foo-test3", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:34:24 +0000 (0:00:00.309) 0:01:07.411 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:34:24 +0000 (0:00:00.039) 0:01:07.451 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:34:24 +0000 (0:00:00.040) 0:01:07.491 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:34:24 +0000 (0:00:00.036) 0:01:07.528 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:34:24 +0000 (0:00:00.024) 0:01:07.552 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:34:24 +0000 (0:00:00.038) 0:01:07.591 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:34:24 +0000 (0:00:00.023) 0:01:07.614 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:34:24 +0000 (0:00:00.502) 0:01:08.117 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:34:24 +0000 (0:00:00.026) 0:01:08.144 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:34:24 +0000 (0:00:00.024) 0:01:08.169 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:34:24 +0000 (0:00:00.053) 0:01:08.222 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:34:24 +0000 (0:00:00.025) 0:01:08.248 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:34:24 +0000 (0:00:00.024) 0:01:08.272 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:34:24 +0000 (0:00:00.022) 0:01:08.294 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:34:24 +0000 (0:00:00.021) 0:01:08.316 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:34:24 +0000 (0:00:00.021) 0:01:08.338 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:34:25 +0000 (0:00:00.053) 0:01:08.391 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:34:25 +0000 (0:00:00.048) 0:01:08.439 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:34:25 +0000 (0:00:00.037) 0:01:08.477 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:34:25 +0000 (0:00:00.037) 0:01:08.514 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:34:25 +0000 (0:00:00.033) 0:01:08.548 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:34:25 +0000 (0:00:00.034) 0:01:08.582 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:34:25 +0000 (0:00:00.035) 0:01:08.618 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:34:25 +0000 (0:00:00.037) 0:01:08.655 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:34:25 +0000 (0:00:00.038) 0:01:08.693 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:34:25 +0000 (0:00:00.036) 0:01:08.729 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:34:25 +0000 (0:00:00.038) 0:01:08.768 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:34:25 +0000 (0:00:00.037) 0:01:08.806 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:34:25 +0000 (0:00:00.035) 0:01:08.841 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:34:25 +0000 (0:00:00.305) 0:01:09.147 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:34:26 +0000 (0:00:00.321) 0:01:09.468 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:34:26 +0000 (0:00:00.050) 0:01:09.519 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:34:26 +0000 (0:00:00.036) 0:01:09.556 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:34:26 +0000 (0:00:00.034) 0:01:09.591 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:34:26 +0000 (0:00:00.034) 0:01:09.625 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:34:26 +0000 (0:00:00.033) 0:01:09.658 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:34:26 +0000 (0:00:00.038) 0:01:09.697 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:34:26 +0000 (0:00:00.035) 0:01:09.732 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:34:26 +0000 (0:00:00.067) 0:01:09.800 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:34:26 +0000 (0:00:00.095) 0:01:09.895 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:34:26 +0000 (0:00:00.057) 0:01:09.952 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test3" ], "delta": "0:00:00.027999", "end": "2022-07-21 14:34:26.834868", "rc": 0, "start": "2022-07-21 14:34:26.806869" } STDOUT: LVM2_LV_NAME=test3 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:34:26 +0000 (0:00:00.346) 0:01:10.298 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:34:26 +0000 (0:00:00.057) 0:01:10.356 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:34:27 +0000 (0:00:00.059) 0:01:10.415 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:34:27 +0000 (0:00:00.043) 0:01:10.458 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:34:27 +0000 (0:00:00.049) 0:01:10.508 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:34:27 +0000 (0:00:00.041) 0:01:10.549 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:34:27 +0000 (0:00:00.036) 0:01:10.586 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 18:34:27 +0000 (0:00:00.037) 0:01:10.624 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 18:34:27 +0000 (0:00:00.021) 0:01:10.645 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Remove two of the LVs] *************************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes.yml:66 Thursday 21 July 2022 18:34:27 +0000 (0:00:00.035) 0:01:10.681 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 18:34:27 +0000 (0:00:00.053) 0:01:10.734 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 18:34:27 +0000 (0:00:00.036) 0:01:10.770 ********* ok: [/cache/rhel-7.qcow2] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 18:34:27 +0000 (0:00:00.434) 0:01:11.205 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 18:34:27 +0000 (0:00:00.065) 0:01:11.271 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 18:34:27 +0000 (0:00:00.036) 0:01:11.308 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 18:34:27 +0000 (0:00:00.038) 0:01:11.346 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 18:34:28 +0000 (0:00:00.060) 0:01:11.406 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 18:34:28 +0000 (0:00:00.022) 0:01:11.429 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 18:34:28 +0000 (0:00:00.029) 0:01:11.458 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "sdb" ], "name": "foo", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "3g" }, { "mount_point": "/opt/test2", "name": "test2", "size": "3g", "state": "absent" }, { "mount_point": "/opt/test3", "name": "test3", "size": "3g", "state": "absent" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 18:34:28 +0000 (0:00:00.041) 0:01:11.500 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 18:34:28 +0000 (0:00:00.039) 0:01:11.539 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 18:34:28 +0000 (0:00:00.030) 0:01:11.569 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 18:34:28 +0000 (0:00:00.030) 0:01:11.600 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 18:34:28 +0000 (0:00:00.034) 0:01:11.634 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 18:34:28 +0000 (0:00:00.029) 0:01:11.664 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 18:34:28 +0000 (0:00:00.109) 0:01:11.774 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 18:34:28 +0000 (0:00:00.059) 0:01:11.833 ********* changed: [/cache/rhel-7.qcow2] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test3", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test3", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/foo-test2", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test2", "fs_type": null } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/mapper/foo-test1", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "absent" }, { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "absent" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 18:34:30 +0000 (0:00:02.341) 0:01:14.174 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 18:34:30 +0000 (0:00:00.043) 0:01:14.218 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 18:34:30 +0000 (0:00:00.023) 0:01:14.241 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test3", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test3", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/foo-test2", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test2", "fs_type": null } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/mapper/foo-test1", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "absent" }, { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "absent" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 18:34:30 +0000 (0:00:00.063) 0:01:14.304 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 18:34:30 +0000 (0:00:00.042) 0:01:14.347 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 18:34:30 +0000 (0:00:00.037) 0:01:14.384 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test3', 'state': 'absent', 'path': '/opt/test3', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "absent" }, "name": "/opt/test3", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test3" } redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test2', 'state': 'absent', 'path': '/opt/test2', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "absent" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 18:34:31 +0000 (0:00:00.650) 0:01:15.035 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 18:34:32 +0000 (0:00:00.474) 0:01:15.509 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 18:34:32 +0000 (0:00:00.333) 0:01:15.843 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 18:34:32 +0000 (0:00:00.458) 0:01:16.302 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658428407.582735, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658201031.524, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 70, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658200515.884, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744071677828413", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 18:34:33 +0000 (0:00:00.301) 0:01:16.604 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 18:34:33 +0000 (0:00:00.025) 0:01:16.630 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes.yml:86 Thursday 21 July 2022 18:34:34 +0000 (0:00:00.841) 0:01:17.471 ********* included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 18:34:34 +0000 (0:00:00.052) 0:01:17.524 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test2", "_mount_id": "/dev/mapper/foo-test2", "_raw_device": "/dev/mapper/foo-test2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "/dev/mapper/foo-test3", "_mount_id": "/dev/mapper/foo-test3", "_raw_device": "/dev/mapper/foo-test3", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 18:34:34 +0000 (0:00:00.106) 0:01:17.630 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 18:34:34 +0000 (0:00:00.037) 0:01:17.668 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "3G", "type": "lvm", "uuid": "fc66144f-ab8b-462b-8161-13a4b348250a" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "fUmW0y-6Mtw-liD0-pqhj-52b5-lflx-H3Il8d" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-18-33-04-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 18:34:34 +0000 (0:00:00.367) 0:01:18.036 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003465", "end": "2022-07-21 14:34:34.926295", "rc": 0, "start": "2022-07-21 14:34:34.922830" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 18:34:34 +0000 (0:00:00.347) 0:01:18.384 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003725", "end": "2022-07-21 14:34:35.232728", "failed_when_result": false, "rc": 0, "start": "2022-07-21 14:34:35.229003" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 18:34:35 +0000 (0:00:00.306) 0:01:18.690 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, {'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, {'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 18:34:35 +0000 (0:00:00.062) 0:01:18.753 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 18:34:35 +0000 (0:00:00.032) 0:01:18.786 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members) included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 18:34:35 +0000 (0:00:00.045) 0:01:18.831 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sdb" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 18:34:35 +0000 (0:00:00.052) 0:01:18.884 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb", "pv": "/dev/sdb" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 18:34:35 +0000 (0:00:00.304) 0:01:19.189 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 18:34:35 +0000 (0:00:00.051) 0:01:19.240 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sdb" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 18:34:35 +0000 (0:00:00.053) 0:01:19.294 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 18:34:35 +0000 (0:00:00.050) 0:01:19.345 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 18:34:35 +0000 (0:00:00.038) 0:01:19.383 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.048) 0:01:19.432 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.022) 0:01:19.454 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.040) 0:01:19.495 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml for /cache/rhel-7.qcow2 TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.040) 0:01:19.536 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.023) 0:01:19.560 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.022) 0:01:19.583 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.022) 0:01:19.605 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.023) 0:01:19.629 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.022) 0:01:19.652 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.022) 0:01:19.675 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.025) 0:01:19.701 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.039) 0:01:19.740 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.043) 0:01:19.784 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.055) 0:01:19.840 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.027) 0:01:19.868 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.026) 0:01:19.895 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.025) 0:01:19.920 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.060) 0:01:19.981 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.028) 0:01:20.009 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.026) 0:01:20.035 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.025) 0:01:20.061 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.027) 0:01:20.089 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.027) 0:01:20.116 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml for /cache/rhel-7.qcow2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.044) 0:01:20.161 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.057) 0:01:20.219 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.025) 0:01:20.244 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.025) 0:01:20.270 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.028) 0:01:20.299 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.027) 0:01:20.326 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.025) 0:01:20.352 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 18:34:36 +0000 (0:00:00.025) 0:01:20.377 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.024) 0:01:20.402 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.024) 0:01:20.426 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.024) 0:01:20.450 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.026) 0:01:20.477 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.024) 0:01:20.501 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.024) 0:01:20.525 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2 TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.047) 0:01:20.573 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.052) 0:01:20.626 ********* skipping: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "_storage_test_pool_member_path": "/dev/sdb", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.029) 0:01:20.656 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/sdb) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.044) 0:01:20.700 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.052) 0:01:20.752 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.053) 0:01:20.806 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.039) 0:01:20.845 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.038) 0:01:20.884 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.037) 0:01:20.922 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.036) 0:01:20.959 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.039) 0:01:20.998 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.050) 0:01:21.049 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.063) 0:01:21.112 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.025) 0:01:21.138 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.025) 0:01:21.163 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.030) 0:01:21.194 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.082) 0:01:21.277 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.030) 0:01:21.307 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.030) 0:01:21.337 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 18:34:37 +0000 (0:00:00.027) 0:01:21.364 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.038) 0:01:21.403 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.026) 0:01:21.430 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.025) 0:01:21.455 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.025) 0:01:21.480 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.026) 0:01:21.507 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.024) 0:01:21.532 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.023) 0:01:21.556 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.023) 0:01:21.579 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.038) 0:01:21.617 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.023) 0:01:21.641 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.023) 0:01:21.664 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.023) 0:01:21.688 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.025) 0:01:21.713 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.023) 0:01:21.737 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.024) 0:01:21.762 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.023) 0:01:21.785 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.036) 0:01:21.821 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.033) 0:01:21.855 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test2', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test2', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test2', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test3', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test3', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test3', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.052) 0:01:21.908 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.048) 0:01:21.957 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.082) 0:01:22.039 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.045) 0:01:22.085 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "fc66144f-ab8b-462b-8161-13a4b348250a" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "fc66144f-ab8b-462b-8161-13a4b348250a" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.060) 0:01:22.146 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.054) 0:01:22.200 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.051) 0:01:22.252 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.053) 0:01:22.306 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.025) 0:01:22.331 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.025) 0:01:22.356 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:34:38 +0000 (0:00:00.025) 0:01:22.382 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:34:39 +0000 (0:00:00.039) 0:01:22.422 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:34:39 +0000 (0:00:00.117) 0:01:22.539 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:34:39 +0000 (0:00:00.098) 0:01:22.637 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:34:39 +0000 (0:00:00.090) 0:01:22.727 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:34:39 +0000 (0:00:00.038) 0:01:22.766 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:34:39 +0000 (0:00:00.076) 0:01:22.842 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:34:39 +0000 (0:00:00.087) 0:01:22.929 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:34:39 +0000 (0:00:00.084) 0:01:23.014 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658428418.179735, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658428418.179735, "dev": 5, "device_type": 64514, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 35068, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658428418.179735, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:34:39 +0000 (0:00:00.363) 0:01:23.378 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:34:40 +0000 (0:00:00.038) 0:01:23.416 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:34:40 +0000 (0:00:00.038) 0:01:23.455 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:34:40 +0000 (0:00:00.036) 0:01:23.491 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:34:40 +0000 (0:00:00.024) 0:01:23.516 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:34:40 +0000 (0:00:00.037) 0:01:23.553 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:34:40 +0000 (0:00:00.023) 0:01:23.577 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:34:40 +0000 (0:00:00.561) 0:01:24.139 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:34:40 +0000 (0:00:00.023) 0:01:24.162 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:34:40 +0000 (0:00:00.023) 0:01:24.186 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:34:40 +0000 (0:00:00.051) 0:01:24.237 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:34:40 +0000 (0:00:00.024) 0:01:24.262 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:34:40 +0000 (0:00:00.024) 0:01:24.287 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:34:40 +0000 (0:00:00.023) 0:01:24.310 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:34:40 +0000 (0:00:00.023) 0:01:24.334 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:34:40 +0000 (0:00:00.023) 0:01:24.357 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:34:41 +0000 (0:00:00.054) 0:01:24.411 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:34:41 +0000 (0:00:00.052) 0:01:24.464 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:34:41 +0000 (0:00:00.039) 0:01:24.503 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:34:41 +0000 (0:00:00.038) 0:01:24.541 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:34:41 +0000 (0:00:00.041) 0:01:24.583 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:34:41 +0000 (0:00:00.033) 0:01:24.617 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:34:41 +0000 (0:00:00.037) 0:01:24.655 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:34:41 +0000 (0:00:00.038) 0:01:24.693 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:34:41 +0000 (0:00:00.044) 0:01:24.738 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:34:41 +0000 (0:00:00.039) 0:01:24.777 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:34:41 +0000 (0:00:00.038) 0:01:24.816 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:34:41 +0000 (0:00:00.040) 0:01:24.856 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:34:41 +0000 (0:00:00.043) 0:01:24.899 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:34:41 +0000 (0:00:00.301) 0:01:25.201 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:34:42 +0000 (0:00:00.321) 0:01:25.522 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:34:42 +0000 (0:00:00.091) 0:01:25.613 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:34:42 +0000 (0:00:00.040) 0:01:25.654 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:34:42 +0000 (0:00:00.036) 0:01:25.691 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:34:42 +0000 (0:00:00.035) 0:01:25.726 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:34:42 +0000 (0:00:00.035) 0:01:25.761 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:34:42 +0000 (0:00:00.036) 0:01:25.798 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:34:42 +0000 (0:00:00.079) 0:01:25.877 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:34:42 +0000 (0:00:00.038) 0:01:25.916 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:34:42 +0000 (0:00:00.037) 0:01:25.953 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:34:42 +0000 (0:00:00.055) 0:01:26.009 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.024759", "end": "2022-07-21 14:34:42.867795", "rc": 0, "start": "2022-07-21 14:34:42.843036" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:34:42 +0000 (0:00:00.319) 0:01:26.328 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:34:42 +0000 (0:00:00.051) 0:01:26.379 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.053) 0:01:26.433 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.044) 0:01:26.477 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.037) 0:01:26.514 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.038) 0:01:26.553 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.037) 0:01:26.590 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.036) 0:01:26.626 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.048) 0:01:26.675 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.079) 0:01:26.754 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.041) 0:01:26.796 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.055) 0:01:26.851 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.023) 0:01:26.875 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.053) 0:01:26.928 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.036) 0:01:26.965 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.024) 0:01:26.989 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.022) 0:01:27.012 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.023) 0:01:27.035 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.035) 0:01:27.071 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.062) 0:01:27.133 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.023) 0:01:27.156 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.049) 0:01:27.206 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.036) 0:01:27.242 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.036) 0:01:27.279 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.023) 0:01:27.302 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:34:43 +0000 (0:00:00.023) 0:01:27.326 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:34:44 +0000 (0:00:00.302) 0:01:27.629 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:34:44 +0000 (0:00:00.084) 0:01:27.714 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:34:44 +0000 (0:00:00.023) 0:01:27.737 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:34:44 +0000 (0:00:00.075) 0:01:27.813 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:34:44 +0000 (0:00:00.060) 0:01:27.873 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:34:44 +0000 (0:00:00.023) 0:01:27.897 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:34:44 +0000 (0:00:00.023) 0:01:27.920 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.508) 0:01:28.429 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.022) 0:01:28.452 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.022) 0:01:28.475 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.022) 0:01:28.498 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.023) 0:01:28.521 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.022) 0:01:28.544 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.022) 0:01:28.566 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.024) 0:01:28.591 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.022) 0:01:28.613 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.049) 0:01:28.663 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.052) 0:01:28.716 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.034) 0:01:28.751 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.033) 0:01:28.785 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.033) 0:01:28.818 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.035) 0:01:28.854 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.034) 0:01:28.889 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.034) 0:01:28.924 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.035) 0:01:28.959 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.036) 0:01:28.995 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.034) 0:01:29.030 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.034) 0:01:29.065 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.035) 0:01:29.100 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.024) 0:01:29.125 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.036) 0:01:29.161 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.035) 0:01:29.197 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.036) 0:01:29.233 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.037) 0:01:29.270 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.035) 0:01:29.305 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.035) 0:01:29.341 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:34:45 +0000 (0:00:00.035) 0:01:29.376 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.036) 0:01:29.413 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.033) 0:01:29.446 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.066) 0:01:29.513 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.022) 0:01:29.536 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.023) 0:01:29.559 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.052) 0:01:29.612 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.023) 0:01:29.636 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.023) 0:01:29.659 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.023) 0:01:29.682 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.023) 0:01:29.706 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.023) 0:01:29.729 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.034) 0:01:29.763 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.049) 0:01:29.813 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.078) 0:01:29.891 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test3" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.038) 0:01:29.930 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.053) 0:01:29.983 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.022) 0:01:30.006 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.046) 0:01:30.053 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.034) 0:01:30.088 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.022) 0:01:30.110 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.023) 0:01:30.134 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.021) 0:01:30.156 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.032) 0:01:30.189 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.067) 0:01:30.256 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.024) 0:01:30.280 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.045) 0:01:30.326 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:34:46 +0000 (0:00:00.033) 0:01:30.360 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:34:47 +0000 (0:00:00.033) 0:01:30.393 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:34:47 +0000 (0:00:00.023) 0:01:30.417 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:34:47 +0000 (0:00:00.021) 0:01:30.438 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:34:47 +0000 (0:00:00.293) 0:01:30.731 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:34:47 +0000 (0:00:00.037) 0:01:30.769 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:34:47 +0000 (0:00:00.023) 0:01:30.792 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:34:47 +0000 (0:00:00.046) 0:01:30.839 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:34:47 +0000 (0:00:00.023) 0:01:30.862 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:34:47 +0000 (0:00:00.023) 0:01:30.886 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:34:47 +0000 (0:00:00.023) 0:01:30.909 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.540) 0:01:31.450 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.055) 0:01:31.505 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.023) 0:01:31.529 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.023) 0:01:31.552 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.023) 0:01:31.575 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.024) 0:01:31.600 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.026) 0:01:31.626 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.026) 0:01:31.653 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.045) 0:01:31.698 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.054) 0:01:31.752 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.048) 0:01:31.801 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.034) 0:01:31.836 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.034) 0:01:31.870 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.038) 0:01:31.908 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.032) 0:01:31.941 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.035) 0:01:31.977 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.037) 0:01:32.014 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.040) 0:01:32.055 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.036) 0:01:32.091 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.036) 0:01:32.128 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.041) 0:01:32.170 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.064) 0:01:32.235 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.026) 0:01:32.262 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.036) 0:01:32.299 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.034) 0:01:32.334 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:34:48 +0000 (0:00:00.036) 0:01:32.370 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.034) 0:01:32.405 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.036) 0:01:32.441 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.036) 0:01:32.478 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.041) 0:01:32.520 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.038) 0:01:32.558 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.037) 0:01:32.596 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.037) 0:01:32.633 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.025) 0:01:32.659 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.024) 0:01:32.683 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.026) 0:01:32.710 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.025) 0:01:32.735 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.024) 0:01:32.760 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.022) 0:01:32.783 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.022) 0:01:32.806 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.023) 0:01:32.829 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.036) 0:01:32.865 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.021) 0:01:32.886 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Re-run the previous role invocation to ensure idempotence] *************** task path: /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes.yml:88 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.074) 0:01:32.961 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.056) 0:01:33.017 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 18:34:49 +0000 (0:00:00.037) 0:01:33.055 ********* ok: [/cache/rhel-7.qcow2] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 18:34:50 +0000 (0:00:00.456) 0:01:33.512 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 18:34:50 +0000 (0:00:00.060) 0:01:33.573 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 18:34:50 +0000 (0:00:00.035) 0:01:33.608 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 18:34:50 +0000 (0:00:00.032) 0:01:33.641 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 18:34:50 +0000 (0:00:00.056) 0:01:33.698 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 18:34:50 +0000 (0:00:00.023) 0:01:33.721 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 18:34:50 +0000 (0:00:00.029) 0:01:33.751 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "sdb" ], "name": "foo", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "3g" }, { "mount_point": "/opt/test2", "name": "test2", "size": "3g", "state": "absent" }, { "mount_point": "/opt/test3", "name": "test3", "size": "3g", "state": "absent" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 18:34:50 +0000 (0:00:00.042) 0:01:33.793 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 18:34:50 +0000 (0:00:00.042) 0:01:33.836 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 18:34:50 +0000 (0:00:00.029) 0:01:33.865 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 18:34:50 +0000 (0:00:00.027) 0:01:33.893 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 18:34:50 +0000 (0:00:00.027) 0:01:33.921 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 18:34:50 +0000 (0:00:00.029) 0:01:33.950 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 18:34:50 +0000 (0:00:00.055) 0:01:34.005 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 18:34:50 +0000 (0:00:00.020) 0:01:34.026 ********* ok: [/cache/rhel-7.qcow2] => { "actions": [], "changed": false, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/mapper/foo-test1", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "", "_mount_id": "", "_raw_device": "", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "", "_mount_id": "", "_raw_device": "", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 18:34:51 +0000 (0:00:01.199) 0:01:35.226 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 18:34:51 +0000 (0:00:00.036) 0:01:35.263 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 18:34:51 +0000 (0:00:00.021) 0:01:35.284 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/mapper/foo-test1", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "", "_mount_id": "", "_raw_device": "", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "", "_mount_id": "", "_raw_device": "", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 18:34:51 +0000 (0:00:00.040) 0:01:35.325 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "", "_mount_id": "", "_raw_device": "", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "", "_mount_id": "", "_raw_device": "", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 18:34:52 +0000 (0:00:00.072) 0:01:35.398 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 18:34:52 +0000 (0:00:00.039) 0:01:35.437 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 18:34:52 +0000 (0:00:00.036) 0:01:35.474 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 18:34:52 +0000 (0:00:00.494) 0:01:35.969 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'dump': 0, 'passno': 0, 'fstype': 'xfs', 'state': 'mounted', 'path': '/opt/test1', 'opts': 'defaults'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 18:34:52 +0000 (0:00:00.381) 0:01:36.350 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 18:34:53 +0000 (0:00:00.535) 0:01:36.885 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658428407.582735, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658201031.524, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 70, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658200515.884, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744071677828413", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 18:34:53 +0000 (0:00:00.320) 0:01:37.206 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 18:34:53 +0000 (0:00:00.022) 0:01:37.228 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes.yml:108 Thursday 21 July 2022 18:34:54 +0000 (0:00:00.859) 0:01:38.087 ********* included: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml for /cache/rhel-7.qcow2 TASK [Print out pool information] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 18:34:54 +0000 (0:00:00.048) 0:01:38.136 ********* ok: [/cache/rhel-7.qcow2] => { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-2", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "", "_mount_id": "", "_raw_device": "", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null }, { "_device": "", "_mount_id": "", "_raw_device": "", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "absent", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 18:34:54 +0000 (0:00:00.057) 0:01:38.194 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 18:34:54 +0000 (0:00:00.037) 0:01:38.232 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "info": { "/dev/fd0": { "fstype": "", "label": "", "name": "/dev/fd0", "size": "4K", "type": "disk", "uuid": "" }, "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "3G", "type": "lvm", "uuid": "fc66144f-ab8b-462b-8161-13a4b348250a" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "fUmW0y-6Mtw-liD0-pqhj-52b5-lflx-H3Il8d" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-18-33-04-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "xfs", "label": "", "name": "/dev/vda1", "size": "10G", "type": "partition", "uuid": "21864ae1-1c29-4009-a1c2-151e41d0e053" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 18:34:56 +0000 (0:00:01.328) 0:01:39.560 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003801", "end": "2022-07-21 14:34:56.417777", "rc": 0, "start": "2022-07-21 14:34:56.413976" } STDOUT: # # /etc/fstab # Created by anaconda on Tue Jul 19 03:15:15 2022 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=21864ae1-1c29-4009-a1c2-151e41d0e053 / xfs defaults 0 0 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 18:34:56 +0000 (0:00:00.317) 0:01:39.877 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003504", "end": "2022-07-21 14:34:56.723225", "failed_when_result": false, "rc": 0, "start": "2022-07-21 14:34:56.719721" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 18:34:56 +0000 (0:00:00.305) 0:01:40.183 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml for /cache/rhel-7.qcow2 => (item={'name': 'foo', 'encryption_password': None, 'state': 'present', 'raid_metadata_version': None, 'encryption': False, 'encryption_key_size': None, 'disks': ['sdb'], 'raid_level': None, 'encryption_luks_version': None, 'raid_device_count': None, 'encryption_key': None, 'volumes': [{'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, {'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}, {'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}], 'raid_chunk_size': None, 'type': 'lvm', 'encryption_cipher': None, 'raid_spare_count': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 18:34:56 +0000 (0:00:00.064) 0:01:40.247 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 18:34:56 +0000 (0:00:00.076) 0:01:40.324 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml for /cache/rhel-7.qcow2 => (item=members) included: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-7.qcow2 => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.098) 0:01:40.422 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sdb" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.061) 0:01:40.483 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb", "pv": "/dev/sdb" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.301) 0:01:40.785 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.049) 0:01:40.835 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sdb" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.048) 0:01:40.883 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.063) 0:01:40.947 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.040) 0:01:40.987 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.048) 0:01:41.036 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.023) 0:01:41.059 ********* ok: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.042) 0:01:41.102 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml for /cache/rhel-7.qcow2 TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.040) 0:01:41.142 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.023) 0:01:41.166 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.024) 0:01:41.190 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.022) 0:01:41.213 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.023) 0:01:41.236 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.023) 0:01:41.259 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.024) 0:01:41.284 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.022) 0:01:41.307 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.032) 0:01:41.340 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-7.qcow2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 18:34:57 +0000 (0:00:00.042) 0:01:41.382 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.051) 0:01:41.434 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.026) 0:01:41.460 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.027) 0:01:41.487 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.026) 0:01:41.513 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.026) 0:01:41.539 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.025) 0:01:41.565 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.026) 0:01:41.592 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.026) 0:01:41.619 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.026) 0:01:41.646 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.026) 0:01:41.672 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml for /cache/rhel-7.qcow2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.044) 0:01:41.717 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.098) 0:01:41.815 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.024) 0:01:41.840 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.023) 0:01:41.864 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.023) 0:01:41.887 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.024) 0:01:41.911 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.026) 0:01:41.937 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.025) 0:01:41.963 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.024) 0:01:41.987 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about thinpool] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.025) 0:01:42.012 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.023) 0:01:42.035 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.028) 0:01:42.064 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.025) 0:01:42.089 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.026) 0:01:42.115 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-7.qcow2 TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.047) 0:01:42.163 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.051) 0:01:42.214 ********* skipping: [/cache/rhel-7.qcow2] => (item=/dev/sdb) => { "_storage_test_pool_member_path": "/dev/sdb", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.028) 0:01:42.243 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-7.qcow2 => (item=/dev/sdb) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.042) 0:01:42.285 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.050) 0:01:42.335 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 18:34:58 +0000 (0:00:00.049) 0:01:42.385 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.036) 0:01:42.421 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.036) 0:01:42.457 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.039) 0:01:42.497 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.034) 0:01:42.531 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.035) 0:01:42.567 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-7.qcow2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.048) 0:01:42.616 ********* included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.059) 0:01:42.675 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.025) 0:01:42.700 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.024) 0:01:42.724 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.023) 0:01:42.748 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.022) 0:01:42.771 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.024) 0:01:42.795 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.022) 0:01:42.818 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.023) 0:01:42.841 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.034) 0:01:42.876 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.024) 0:01:42.900 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.024) 0:01:42.925 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.023) 0:01:42.948 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.024) 0:01:42.972 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.059) 0:01:43.032 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.024) 0:01:43.056 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.023) 0:01:43.079 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.034) 0:01:43.114 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.024) 0:01:43.138 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.023) 0:01:43.161 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.026) 0:01:43.188 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.024) 0:01:43.212 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.024) 0:01:43.236 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.025) 0:01:43.262 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.025) 0:01:43.287 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.035) 0:01:43.323 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 18:34:59 +0000 (0:00:00.036) 0:01:43.360 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '/dev/mapper/foo-test1', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '/dev/mapper/foo-test1', 'size': '3g', 'mount_point': '/opt/test1', 'compression': None, 'encryption_password': None, '_kernel_device': '/dev/dm-2', 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test1', 'raid_device_count': None, 'state': 'present', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '/dev/mapper/foo-test1', 'raid_spare_count': None, 'raid_disks': [], '_raw_kernel_device': '/dev/dm-2', 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': ['sdb'], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test2', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test2', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml for /cache/rhel-7.qcow2 => (item={'_raw_device': '', 'raid_metadata_version': None, 'raid_level': None, 'fs_type': 'xfs', 'mount_options': 'defaults', 'type': 'lvm', '_device': '', 'size': '3g', 'mount_point': '/opt/test3', 'compression': None, 'encryption_password': None, 'encryption': False, 'mount_device_identifier': 'uuid', 'name': 'test3', 'raid_device_count': None, 'state': 'absent', 'vdo_pool_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'encryption_key_size': None, 'encryption_cipher': None, 'encryption_key': None, 'fs_label': '', 'encryption_luks_version': None, 'cache_size': 0, '_mount_id': '', 'raid_spare_count': None, 'raid_disks': [], 'cache_mode': None, 'deduplication': None, 'cached': False, 'fs_overwrite_existing': True, 'disks': [], 'thin': False, 'mount_check': 0, 'mount_passno': 0, 'raid_chunk_size': None, 'cache_devices': [], 'fs_create_options': ''}) TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.050) 0:01:43.410 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.052) 0:01:43.462 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.079) 0:01:43.542 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.039) 0:01:43.581 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "fc66144f-ab8b-462b-8161-13a4b348250a" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 775624, "block_size": 4096, "block_total": 783872, "block_used": 8248, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3176955904, "size_total": 3210739712, "uuid": "fc66144f-ab8b-462b-8161-13a4b348250a" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.056) 0:01:43.638 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.052) 0:01:43.691 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.048) 0:01:43.739 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.052) 0:01:43.792 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.025) 0:01:43.817 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.026) 0:01:43.843 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.025) 0:01:43.868 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.038) 0:01:43.907 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.065) 0:01:43.973 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.051) 0:01:44.024 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.051) 0:01:44.075 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.037) 0:01:44.113 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.038) 0:01:44.152 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.040) 0:01:44.192 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:35:00 +0000 (0:00:00.042) 0:01:44.234 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658428418.179735, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658428418.179735, "dev": 5, "device_type": 64514, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 35068, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658428418.179735, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:35:01 +0000 (0:00:00.316) 0:01:44.550 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:35:01 +0000 (0:00:00.113) 0:01:44.664 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:35:01 +0000 (0:00:00.039) 0:01:44.704 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:35:01 +0000 (0:00:00.039) 0:01:44.743 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:35:01 +0000 (0:00:00.025) 0:01:44.769 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:35:01 +0000 (0:00:00.042) 0:01:44.811 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:35:01 +0000 (0:00:00.026) 0:01:44.838 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:35:01 +0000 (0:00:00.527) 0:01:45.366 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.025) 0:01:45.391 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.023) 0:01:45.415 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.053) 0:01:45.469 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.026) 0:01:45.495 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.025) 0:01:45.521 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.023) 0:01:45.544 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.024) 0:01:45.569 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.023) 0:01:45.592 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.049) 0:01:45.641 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.047) 0:01:45.689 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.037) 0:01:45.727 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.035) 0:01:45.762 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.042) 0:01:45.804 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.038) 0:01:45.842 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.042) 0:01:45.885 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.039) 0:01:45.924 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.038) 0:01:45.963 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.038) 0:01:46.001 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.041) 0:01:46.043 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.038) 0:01:46.082 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:35:02 +0000 (0:00:00.038) 0:01:46.121 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:35:03 +0000 (0:00:00.313) 0:01:46.435 ********* ok: [/cache/rhel-7.qcow2] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:35:03 +0000 (0:00:00.318) 0:01:46.754 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:35:03 +0000 (0:00:00.051) 0:01:46.805 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:35:03 +0000 (0:00:00.038) 0:01:46.844 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:35:03 +0000 (0:00:00.039) 0:01:46.883 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:35:03 +0000 (0:00:00.043) 0:01:46.927 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:35:03 +0000 (0:00:00.038) 0:01:46.966 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:35:03 +0000 (0:00:00.038) 0:01:47.005 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:35:03 +0000 (0:00:00.038) 0:01:47.044 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:35:03 +0000 (0:00:00.041) 0:01:47.085 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:35:03 +0000 (0:00:00.037) 0:01:47.123 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:35:03 +0000 (0:00:00.093) 0:01:47.216 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.023216", "end": "2022-07-21 14:35:04.127083", "rc": 0, "start": "2022-07-21 14:35:04.103867" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:35:04 +0000 (0:00:00.371) 0:01:47.587 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:35:04 +0000 (0:00:00.092) 0:01:47.680 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:35:04 +0000 (0:00:00.094) 0:01:47.775 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:35:04 +0000 (0:00:00.100) 0:01:47.875 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:35:04 +0000 (0:00:00.039) 0:01:47.915 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:35:04 +0000 (0:00:00.039) 0:01:47.955 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:35:04 +0000 (0:00:00.040) 0:01:47.995 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:35:04 +0000 (0:00:00.035) 0:01:48.031 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:35:04 +0000 (0:00:00.052) 0:01:48.083 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:35:04 +0000 (0:00:00.081) 0:01:48.165 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:35:04 +0000 (0:00:00.043) 0:01:48.209 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:35:04 +0000 (0:00:00.056) 0:01:48.265 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:35:04 +0000 (0:00:00.025) 0:01:48.291 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:35:04 +0000 (0:00:00.052) 0:01:48.343 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:35:04 +0000 (0:00:00.042) 0:01:48.386 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.025) 0:01:48.411 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.024) 0:01:48.435 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.023) 0:01:48.459 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.036) 0:01:48.495 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.064) 0:01:48.559 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.023) 0:01:48.583 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.050) 0:01:48.634 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.035) 0:01:48.669 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.033) 0:01:48.703 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.022) 0:01:48.726 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.023) 0:01:48.750 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.301) 0:01:49.051 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.039) 0:01:49.090 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.023) 0:01:49.114 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.036) 0:01:49.151 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.025) 0:01:49.176 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.023) 0:01:49.199 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:35:05 +0000 (0:00:00.023) 0:01:49.223 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.542) 0:01:49.765 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.026) 0:01:49.792 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.026) 0:01:49.819 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.026) 0:01:49.846 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.026) 0:01:49.872 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.025) 0:01:49.898 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.025) 0:01:49.924 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.027) 0:01:49.951 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.025) 0:01:49.977 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.053) 0:01:50.031 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.051) 0:01:50.082 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.041) 0:01:50.124 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.044) 0:01:50.169 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.039) 0:01:50.208 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.039) 0:01:50.248 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.043) 0:01:50.292 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.040) 0:01:50.332 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:35:06 +0000 (0:00:00.039) 0:01:50.371 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.038) 0:01:50.410 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.040) 0:01:50.451 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.037) 0:01:50.488 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.036) 0:01:50.524 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.024) 0:01:50.548 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.039) 0:01:50.588 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.035) 0:01:50.623 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.035) 0:01:50.659 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.035) 0:01:50.694 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.040) 0:01:50.734 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.035) 0:01:50.770 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.035) 0:01:50.805 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.034) 0:01:50.840 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.037) 0:01:50.878 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.034) 0:01:50.912 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.022) 0:01:50.935 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.022) 0:01:50.957 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.023) 0:01:50.981 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.022) 0:01:51.003 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.022) 0:01:51.026 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.022) 0:01:51.048 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.022) 0:01:51.071 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.021) 0:01:51.093 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.065) 0:01:51.158 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.049) 0:01:51.207 ********* included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml for /cache/rhel-7.qcow2 => (item=mount) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-7.qcow2 => (item=fstab) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml for /cache/rhel-7.qcow2 => (item=fs) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml for /cache/rhel-7.qcow2 => (item=device) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-7.qcow2 => (item=encryption) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml for /cache/rhel-7.qcow2 => (item=md) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml for /cache/rhel-7.qcow2 => (item=size) included: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml for /cache/rhel-7.qcow2 => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.110) 0:01:51.318 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_device_path": "" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 18:35:07 +0000 (0:00:00.040) 0:01:51.358 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.053) 0:01:51.412 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.023) 0:01:51.436 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.049) 0:01:51.486 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.035) 0:01:51.521 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.023) 0:01:51.544 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.023) 0:01:51.567 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.026) 0:01:51.594 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.036) 0:01:51.630 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.064) 0:01:51.694 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.026) 0:01:51.721 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.052) 0:01:51.773 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.037) 0:01:51.811 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.038) 0:01:51.849 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.027) 0:01:51.876 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.025) 0:01:51.902 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.309) 0:01:52.212 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.040) 0:01:52.252 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.025) 0:01:52.278 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.040) 0:01:52.318 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.025) 0:01:52.344 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 18:35:08 +0000 (0:00:00.025) 0:01:52.369 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.025) 0:01:52.395 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.511) 0:01:52.907 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.025) 0:01:52.932 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.024) 0:01:52.957 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.023) 0:01:52.981 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.025) 0:01:53.006 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.025) 0:01:53.032 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.022) 0:01:53.054 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.022) 0:01:53.077 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.022) 0:01:53.099 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.051) 0:01:53.150 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.046) 0:01:53.197 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.034) 0:01:53.231 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.034) 0:01:53.265 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.037) 0:01:53.303 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.036) 0:01:53.340 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 18:35:09 +0000 (0:00:00.037) 0:01:53.378 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.047) 0:01:53.425 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.096) 0:01:53.522 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.039) 0:01:53.561 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.039) 0:01:53.601 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.041) 0:01:53.642 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.036) 0:01:53.678 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.023) 0:01:53.702 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.035) 0:01:53.737 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.036) 0:01:53.774 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.033) 0:01:53.807 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.035) 0:01:53.842 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.036) 0:01:53.879 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.037) 0:01:53.916 ********* skipping: [/cache/rhel-7.qcow2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.038) 0:01:53.954 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.039) 0:01:53.994 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [debug] ******************************************************************* task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.038) 0:01:54.032 ********* ok: [/cache/rhel-7.qcow2] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.047) 0:01:54.079 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.025) 0:01:54.104 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.023) 0:01:54.128 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check segment type] ****************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.023) 0:01:54.151 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.024) 0:01:54.176 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.023) 0:01:54.200 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.022) 0:01:54.223 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.022) 0:01:54.245 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmptomayb7j/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.023) 0:01:54.269 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.034) 0:01:54.303 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.021) 0:01:54.325 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Clean up] **************************************************************** task path: /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes.yml:110 Thursday 21 July 2022 18:35:10 +0000 (0:00:00.034) 0:01:54.360 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 18:35:11 +0000 (0:00:00.063) 0:01:54.424 ********* included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 18:35:11 +0000 (0:00:00.036) 0:01:54.461 ********* ok: [/cache/rhel-7.qcow2] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 18:35:11 +0000 (0:00:00.415) 0:01:54.876 ********* skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-7.qcow2] => (item=RedHat_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.yml" } skipping: [/cache/rhel-7.qcow2] => (item=RedHat_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 18:35:11 +0000 (0:00:00.065) 0:01:54.942 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 18:35:11 +0000 (0:00:00.037) 0:01:54.979 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 18:35:11 +0000 (0:00:00.085) 0:01:55.065 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-7.qcow2 TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 18:35:11 +0000 (0:00:00.096) 0:01:55.161 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 18:35:11 +0000 (0:00:00.022) 0:01:55.184 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 18:35:11 +0000 (0:00:00.035) 0:01:55.220 ********* ok: [/cache/rhel-7.qcow2] => { "storage_pools": [ { "disks": [ "sdb" ], "name": "foo", "state": "absent", "volumes": [] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 18:35:11 +0000 (0:00:00.042) 0:01:55.262 ********* ok: [/cache/rhel-7.qcow2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 18:35:11 +0000 (0:00:00.039) 0:01:55.302 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 18:35:11 +0000 (0:00:00.030) 0:01:55.333 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 18:35:11 +0000 (0:00:00.030) 0:01:55.364 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 18:35:12 +0000 (0:00:00.034) 0:01:55.399 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 18:35:12 +0000 (0:00:00.030) 0:01:55.429 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 18:35:12 +0000 (0:00:00.060) 0:01:55.490 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 18:35:12 +0000 (0:00:00.023) 0:01:55.514 ********* changed: [/cache/rhel-7.qcow2] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdb", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "xfsprogs" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 18:35:13 +0000 (0:00:01.556) 0:01:57.070 ********* skipping: [/cache/rhel-7.qcow2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 18:35:13 +0000 (0:00:00.038) 0:01:57.108 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 18:35:13 +0000 (0:00:00.023) 0:01:57.132 ********* ok: [/cache/rhel-7.qcow2] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdb", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb", "/dev/vdc", "/dev/vdd" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "xfsprogs" ], "pools": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 18:35:13 +0000 (0:00:00.043) 0:01:57.176 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 18:35:13 +0000 (0:00:00.036) 0:01:57.212 ********* ok: [/cache/rhel-7.qcow2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 18:35:13 +0000 (0:00:00.035) 0:01:57.247 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-7.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'state': 'absent', 'path': '/opt/test1', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 18:35:14 +0000 (0:00:00.346) 0:01:57.593 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 18:35:14 +0000 (0:00:00.449) 0:01:58.042 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 18:35:14 +0000 (0:00:00.037) 0:01:58.080 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 18:35:15 +0000 (0:00:00.430) 0:01:58.511 ********* ok: [/cache/rhel-7.qcow2] => { "changed": false, "stat": { "atime": 1658428407.582735, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1658201031.524, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 70, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1658200515.884, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "18446744071677828413", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 18:35:15 +0000 (0:00:00.314) 0:01:58.825 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 18:35:15 +0000 (0:00:00.023) 0:01:58.848 ********* ok: [/cache/rhel-7.qcow2] META: role_complete for /cache/rhel-7.qcow2 META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* /cache/rhel-7.qcow2 : ok=748 changed=9 unreachable=0 failed=0 skipped=735 rescued=0 ignored=0 Thursday 21 July 2022 18:35:16 +0000 (0:00:00.902) 0:01:59.751 ********* =============================================================================== fedora.linux_system_roles.storage : make sure blivet is available ------- 9.37s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 2.34s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.87s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : Update facts ------------------------ 1.86s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.72s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.56s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Ensure cryptsetup is present -------------------------------------------- 1.55s /tmp/tmptomayb7j/tests/storage/test-verify-volume-encryption.yml:10 ----------- Collect info about the volumes. ----------------------------------------- 1.45s /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14 --------------------- set up internal repositories -------------------------------------------- 1.42s /cache/rhel-7_setup.yml:5 ----------------------------------------------------- Collect info about the volumes. ----------------------------------------- 1.33s /tmp/tmptomayb7j/tests/storage/verify-role-results.yml:14 --------------------- fedora.linux_system_roles.storage : get service facts ------------------- 1.21s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 1.20s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : set up new/current mounts ----------- 1.18s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Gathering Facts --------------------------------------------------------- 1.00s /tmp/tmptomayb7j/tests/storage/tests_lvm_one_disk_multiple_volumes_scsi_generated.yml:3 fedora.linux_system_roles.storage : set up new/current mounts ----------- 0.94s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 fedora.linux_system_roles.storage : Update facts ------------------------ 0.90s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 fedora.linux_system_roles.storage : Update facts ------------------------ 0.88s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 fedora.linux_system_roles.storage : Update facts ------------------------ 0.86s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 fedora.linux_system_roles.storage : Update facts ------------------------ 0.84s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 fedora.linux_system_roles.storage : Update facts ------------------------ 0.82s /tmp/tmp5bkr4li_/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183