+ cd /tmp/tmpfvho35gc/tests; TEST_SUBJECTS=/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TEST_ARTIFACTS=/tmp/linux-system-role-test-work-pull-linux-system-roles_storage-25-db52957-rhel-8-df0ousve/artifacts ansible-playbook -vv --inventory=/usr/share/ansible/inventory/standard-inventory-qcow2 /tmp/tmpfvho35gc/_setup.yml /tmp/tmpfvho35gc/tests/tests_change_disk_fs.yml ansible-playbook 2.8.0 config file = /home/till/.ansible.cfg configured module search path = ['/home/till/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 3.7.3 (default, May 11 2019, 00:38:04) [GCC 9.1.1 20190503 (Red Hat 9.1.1-1)] Using /home/till/.ansible.cfg as config file PLAYBOOK: _setup.yml *************************************************************************************************************************************************************************************** 2 plays in /tmp/tmpfvho35gc/_setup.yml PLAY [Fail when only localhost is available] *************************************************************************************************************************************************************** META: ran handlers TASK [debug] *********************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/_setup.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "groups": { "all": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "localhost": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "subjects": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "ungrouped": [] } } TASK [fail] ************************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:7 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} META: ran handlers META: ran handlers PLAY [Setup repos] ***************************************************************************************************************************************************************************************** META: ran handlers TASK [set up internal repositories] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:16 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true} META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=2 changed=1 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 PLAYBOOK: tests_change_disk_fs.yml ************************************************************************************************************************************************************************* 1 plays in /tmp/tmpfvho35gc/tests/tests_change_disk_fs.yml PLAY [all] ************************************************************************************************************************************************************************************************* META: ran handlers META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=2 changed=1 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 + cd /tmp/tmpfvho35gc/tests; TEST_SUBJECTS=/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TEST_ARTIFACTS=/tmp/linux-system-role-test-work-pull-linux-system-roles_storage-25-db52957-rhel-8-df0ousve/artifacts ansible-playbook -vv --inventory=/usr/share/ansible/inventory/standard-inventory-qcow2 /tmp/tmpfvho35gc/_setup.yml /tmp/tmpfvho35gc/tests/tests_change_disk_mount.yml ansible-playbook 2.8.0 config file = /home/till/.ansible.cfg configured module search path = ['/home/till/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 3.7.3 (default, May 11 2019, 00:38:04) [GCC 9.1.1 20190503 (Red Hat 9.1.1-1)] Using /home/till/.ansible.cfg as config file PLAYBOOK: _setup.yml *************************************************************************************************************************************************************************************** 2 plays in /tmp/tmpfvho35gc/_setup.yml PLAY [Fail when only localhost is available] *************************************************************************************************************************************************************** META: ran handlers TASK [debug] *********************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/_setup.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "groups": { "all": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "localhost": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "subjects": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "ungrouped": [] } } TASK [fail] ************************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:7 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} META: ran handlers META: ran handlers PLAY [Setup repos] ***************************************************************************************************************************************************************************************** META: ran handlers TASK [set up internal repositories] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:16 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true} META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=2 changed=1 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 PLAYBOOK: tests_change_disk_mount.yml ********************************************************************************************************************************************************************** 1 plays in /tmp/tmpfvho35gc/tests/tests_change_disk_mount.yml PLAY [all] ************************************************************************************************************************************************************************************************* META: ran handlers META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=2 changed=1 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 + cd /tmp/tmpfvho35gc/tests; TEST_SUBJECTS=/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TEST_ARTIFACTS=/tmp/linux-system-role-test-work-pull-linux-system-roles_storage-25-db52957-rhel-8-df0ousve/artifacts ansible-playbook -vv --inventory=/usr/share/ansible/inventory/standard-inventory-qcow2 /tmp/tmpfvho35gc/_setup.yml /tmp/tmpfvho35gc/tests/tests_change_fs.yml ansible-playbook 2.8.0 config file = /home/till/.ansible.cfg configured module search path = ['/home/till/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 3.7.3 (default, May 11 2019, 00:38:04) [GCC 9.1.1 20190503 (Red Hat 9.1.1-1)] Using /home/till/.ansible.cfg as config file PLAYBOOK: _setup.yml *************************************************************************************************************************************************************************************** 2 plays in /tmp/tmpfvho35gc/_setup.yml PLAY [Fail when only localhost is available] *************************************************************************************************************************************************************** META: ran handlers TASK [debug] *********************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/_setup.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "groups": { "all": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "localhost": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "subjects": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "ungrouped": [] } } TASK [fail] ************************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:7 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} META: ran handlers META: ran handlers PLAY [Setup repos] ***************************************************************************************************************************************************************************************** META: ran handlers TASK [set up internal repositories] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:16 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true} META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=2 changed=1 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 PLAYBOOK: tests_change_fs.yml ****************************************************************************************************************************************************************************** 1 plays in /tmp/tmpfvho35gc/tests/tests_change_fs.yml PLAY [all] ************************************************************************************************************************************************************************************************* TASK [Gathering Facts] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/tests_change_fs.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_all_ipv4_addresses": ["10.0.2.15"], "ansible_all_ipv6_addresses": ["fec0::5054:ff:fe12:3456", "fe80::5054:ff:fe12:3456"], "ansible_apparmor": {"status": "disabled"}, "ansible_architecture": "x86_64", "ansible_bios_date": "04/01/2014", "ansible_bios_version": "1.12.0-2.fc30", "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": "ttyS0,115200n8", "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_date_time": {"date": "2019-05-26", "day": "26", "epoch": "1558866235", "hour": "06", "iso8601": "2019-05-26T10:23:55Z", "iso8601_basic": "20190526T062355613376", "iso8601_basic_short": "20190526T062355", "iso8601_micro": "2019-05-26T10:23:55.617580Z", "minute": "23", "month": "05", "second": "55", "time": "06:23:55", "tz": "EDT", "tz_offset": "-0400", "weekday": "Sonntag", "weekday_number": "0", "weeknumber": "20", "year": "2019"}, "ansible_default_ipv4": {"address": "10.0.2.15", "alias": "eth0", "broadcast": "10.0.2.255", "gateway": "10.0.2.2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "netmask": "255.255.255.0", "network": "10.0.2.0", "type": "ether"}, "ansible_default_ipv6": {"address": "fec0::5054:ff:fe12:3456", "gateway": "fe80::2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "prefix": "64", "scope": "site", "type": "ether"}, "ansible_device_links": {"ids": {"sr0": ["ata-QEMU_DVD-ROM_QM00003"]}, "labels": {"sr0": ["cidata"]}, "masters": {}, "uuids": {"sr0": ["2019-05-26-12-22-10-00"], "vda1": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}}, "ansible_devices": {"sr0": {"holders": [], "host": "IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]", "links": {"ids": ["ata-QEMU_DVD-ROM_QM00003"], "labels": ["cidata"], "masters": [], "uuids": ["2019-05-26-12-22-10-00"]}, "model": "QEMU DVD-ROM", "partitions": {}, "removable": "1", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "728", "sectorsize": "2048", "size": "364.00 KB", "support_discard": "0", "vendor": "QEMU", "virtual": 1}, "vda": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {"vda1": {"holders": [], "links": {"ids": [], "labels": [], "masters": [], "uuids": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}, "sectors": "20969439", "sectorsize": 512, "size": "10.00 GB", "start": "2048", "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdb": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdc": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}}, "ansible_distribution": "RedHat", "ansible_distribution_file_parsed": true, "ansible_distribution_file_path": "/etc/redhat-release", "ansible_distribution_file_search_string": "Red Hat", "ansible_distribution_file_variety": "RedHat", "ansible_distribution_major_version": "8", "ansible_distribution_release": "Ootpa", "ansible_distribution_version": "8.0", "ansible_dns": {"nameservers": ["10.0.2.3"], "search": ["virt.pnr.lab.eng.rdu2.redhat.com"]}, "ansible_domain": "virt.pnr.lab.eng.rdu2.redhat.com", "ansible_effective_group_id": 0, "ansible_effective_user_id": 0, "ansible_env": {"DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "HOME": "/root", "LANG": "de_DE.UTF-8", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "LOGNAME": "root", "LS_COLORS": "rs=0:di=38;5;33:ln=38;5;51:mh=00:pi=40;38;5;11:so=38;5;13:do=38;5;5:bd=48;5;232;38;5;11:cd=48;5;232;38;5;3:or=48;5;232;38;5;9:mi=01;05;37;41:su=48;5;196;38;5;15:sg=48;5;11;38;5;16:ca=48;5;196;38;5;226:tw=48;5;10;38;5;16:ow=48;5;10;38;5;21:st=48;5;21;38;5;15:ex=38;5;40:*.tar=38;5;9:*.tgz=38;5;9:*.arc=38;5;9:*.arj=38;5;9:*.taz=38;5;9:*.lha=38;5;9:*.lz4=38;5;9:*.lzh=38;5;9:*.lzma=38;5;9:*.tlz=38;5;9:*.txz=38;5;9:*.tzo=38;5;9:*.t7z=38;5;9:*.zip=38;5;9:*.z=38;5;9:*.dz=38;5;9:*.gz=38;5;9:*.lrz=38;5;9:*.lz=38;5;9:*.lzo=38;5;9:*.xz=38;5;9:*.zst=38;5;9:*.tzst=38;5;9:*.bz2=38;5;9:*.bz=38;5;9:*.tbz=38;5;9:*.tbz2=38;5;9:*.tz=38;5;9:*.deb=38;5;9:*.rpm=38;5;9:*.jar=38;5;9:*.war=38;5;9:*.ear=38;5;9:*.sar=38;5;9:*.rar=38;5;9:*.alz=38;5;9:*.ace=38;5;9:*.zoo=38;5;9:*.cpio=38;5;9:*.7z=38;5;9:*.rz=38;5;9:*.cab=38;5;9:*.wim=38;5;9:*.swm=38;5;9:*.dwm=38;5;9:*.esd=38;5;9:*.jpg=38;5;13:*.jpeg=38;5;13:*.mjpg=38;5;13:*.mjpeg=38;5;13:*.gif=38;5;13:*.bmp=38;5;13:*.pbm=38;5;13:*.pgm=38;5;13:*.ppm=38;5;13:*.tga=38;5;13:*.xbm=38;5;13:*.xpm=38;5;13:*.tif=38;5;13:*.tiff=38;5;13:*.png=38;5;13:*.svg=38;5;13:*.svgz=38;5;13:*.mng=38;5;13:*.pcx=38;5;13:*.mov=38;5;13:*.mpg=38;5;13:*.mpeg=38;5;13:*.m2v=38;5;13:*.mkv=38;5;13:*.webm=38;5;13:*.ogm=38;5;13:*.mp4=38;5;13:*.m4v=38;5;13:*.mp4v=38;5;13:*.vob=38;5;13:*.qt=38;5;13:*.nuv=38;5;13:*.wmv=38;5;13:*.asf=38;5;13:*.rm=38;5;13:*.rmvb=38;5;13:*.flc=38;5;13:*.avi=38;5;13:*.fli=38;5;13:*.flv=38;5;13:*.gl=38;5;13:*.dl=38;5;13:*.xcf=38;5;13:*.xwd=38;5;13:*.yuv=38;5;13:*.cgm=38;5;13:*.emf=38;5;13:*.ogv=38;5;13:*.ogx=38;5;13:*.aac=38;5;45:*.au=38;5;45:*.flac=38;5;45:*.m4a=38;5;45:*.mid=38;5;45:*.midi=38;5;45:*.mka=38;5;45:*.mp3=38;5;45:*.mpc=38;5;45:*.ogg=38;5;45:*.ra=38;5;45:*.wav=38;5;45:*.oga=38;5;45:*.opus=38;5;45:*.spx=38;5;45:*.xspf=38;5;45:", "MAIL": "/var/mail/root", "PATH": "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "PWD": "/root", "SELINUX_LEVEL_REQUESTED": "", "SELINUX_ROLE_REQUESTED": "", "SELINUX_USE_CURRENT_RANGE": "", "SHELL": "/bin/bash", "SHLVL": "2", "SSH_CLIENT": "10.0.2.2 33116 22", "SSH_CONNECTION": "10.0.2.2 33116 10.0.2.15 22", "SSH_TTY": "/dev/pts/0", "TERM": "xterm-256color", "USER": "root", "XDG_RUNTIME_DIR": "/run/user/0", "XDG_SESSION_ID": "1", "_": "/usr/libexec/platform-python"}, "ansible_eth0": {"active": true, "device": "eth0", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "off [requested on]", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "off [fixed]", "netns_local": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "off [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "on [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "off", "tcp_segmentation_offload": "off", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "tx_checksumming": "off", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "off [fixed]", "tx_nocache_copy": "off", "tx_scatter_gather": "off [fixed]", "tx_scatter_gather_fraglist": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_tcp6_segmentation": "off [fixed]", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off [fixed]", "tx_tcp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "off [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "10.0.2.15", "broadcast": "10.0.2.255", "netmask": "255.255.255.0", "network": "10.0.2.0"}, "ipv6": [{"address": "fec0::5054:ff:fe12:3456", "prefix": "64", "scope": "site"}, {"address": "fe80::5054:ff:fe12:3456", "prefix": "64", "scope": "link"}], "macaddress": "52:54:00:12:34:56", "module": "virtio_net", "mtu": 1500, "pciid": "virtio0", "promisc": false, "speed": -1, "timestamping": ["tx_software", "rx_software", "software"], "type": "ether"}, "ansible_fibre_channel_wwn": [], "ansible_fips": false, "ansible_form_factor": "Other", "ansible_fqdn": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_hostname": "ibm-p8-kvm-03-guest-02", "ansible_hostnqn": "", "ansible_interfaces": ["lo", "eth0"], "ansible_is_chroot": false, "ansible_iscsi_iqn": "", "ansible_kernel": "4.18.0-80.el8.x86_64", "ansible_lo": {"active": true, "device": "lo", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "on", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "on [fixed]", "netns_local": "on [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "on [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "off [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "on", "tcp_segmentation_offload": "on", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "tx_checksumming": "on", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "on [fixed]", "tx_nocache_copy": "off [fixed]", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tx_sctp_segmentation": "on", "tx_tcp6_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp_segmentation": "on", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "on [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "127.0.0.1", "broadcast": "host", "netmask": "255.0.0.0", "network": "127.0.0.0"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "mtu": 65536, "promisc": false, "timestamping": ["tx_software", "rx_software", "software"], "type": "loopback"}, "ansible_local": {}, "ansible_lsb": {}, "ansible_machine": "x86_64", "ansible_machine_id": "cba1be223eba4c078f6aa69d5d727df8", "ansible_memfree_mb": 357, "ansible_memory_mb": {"nocache": {"free": 595, "used": 225}, "real": {"free": 357, "total": 820, "used": 463}, "swap": {"cached": 0, "free": 0, "total": 0, "used": 0}}, "ansible_memtotal_mb": 820, "ansible_mounts": [{"block_available": 2290069, "block_size": 4096, "block_total": 2618619, "block_used": 328550, "device": "/dev/vda1", "fstype": "xfs", "inode_available": 5207720, "inode_total": 5242352, "inode_used": 34632, "mount": "/", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 9380122624, "size_total": 10725863424, "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}], "ansible_nodename": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_os_family": "RedHat", "ansible_pkg_mgr": "dnf", "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": ["ttyS0", "ttyS0,115200n8"], "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_processor": ["0", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "1", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "2", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz"], "ansible_processor_cores": 1, "ansible_processor_count": 3, "ansible_processor_threads_per_core": 1, "ansible_processor_vcpus": 3, "ansible_product_name": "Standard PC (i440FX + PIIX, 1996)", "ansible_product_serial": "NA", "ansible_product_uuid": "NA", "ansible_product_version": "pc-i440fx-3.1", "ansible_python": {"executable": "/usr/libexec/platform-python", "has_sslcontext": true, "type": "cpython", "version": {"major": 3, "micro": 8, "minor": 6, "releaselevel": "final", "serial": 0}, "version_info": [3, 6, 8, "final", 0]}, "ansible_python_version": "3.6.8", "ansible_real_group_id": 0, "ansible_real_user_id": 0, "ansible_selinux": {"config_mode": "enforcing", "mode": "enforcing", "policyvers": 31, "status": "enabled", "type": "targeted"}, "ansible_selinux_python_present": true, "ansible_service_mgr": "systemd", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNepBea4g9eV/eGHPP/SQsFC27+DquW8imAs/3tU0m+7pmy8Bv+nfWkP8zJNTCqhii7h5r6A7Nfyaf7oyppGWBk=", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIPvQyPw/sTd+2zVaVH0pt5MpMXvUkH73unVFw5ToUlXN", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABAQDhHcMp7yhmX8vYAeBZEi0Mb2c6WW5ZtmesO6zhmaN/oY7OJ6t2el24TKSGvv+9KAlVraWTKGtbvLwPKOPWzr4ClhDS4ym5/lfyycEJhd0ymLajgAVrMQAKtt0639rf1gR4Yj2GYAwNUb/EvzApkuuftWBBhi57TXc+1HgomH2mAzo54w4mW+3zJUEa5igTXBqEse4QcTf9lNZ6cv1eUG7Xl2jyhDA1ngJEU3Dz3QrJ0mfhc2O5n7dTLfkc69PisdfXjw22hEup3IHDU4eB/KiJM9nJms3vdUZK8H79AhF7YAGNK2OxbF9FM6RfuYmgFUH0trQcepljdG1ENMt+cLHd", "ansible_swapfree_mb": 0, "ansible_swaptotal_mb": 0, "ansible_system": "Linux", "ansible_system_capabilities": ["cap_chown", "cap_dac_override", "cap_dac_read_search", "cap_fowner", "cap_fsetid", "cap_kill", "cap_setgid", "cap_setuid", "cap_setpcap", "cap_linux_immutable", "cap_net_bind_service", "cap_net_broadcast", "cap_net_admin", "cap_net_raw", "cap_ipc_lock", "cap_ipc_owner", "cap_sys_module", "cap_sys_rawio", "cap_sys_chroot", "cap_sys_ptrace", "cap_sys_pacct", "cap_sys_admin", "cap_sys_boot", "cap_sys_nice", "cap_sys_resource", "cap_sys_time", "cap_sys_tty_config", "cap_mknod", "cap_lease", "cap_audit_write", "cap_audit_control", "cap_setfcap", "cap_mac_override", "cap_mac_admin", "cap_syslog", "cap_wake_alarm", "cap_block_suspend", "cap_audit_read+ep"], "ansible_system_capabilities_enforced": "True", "ansible_system_vendor": "QEMU", "ansible_uptime_seconds": 100, "ansible_user_dir": "/root", "ansible_user_gecos": "root", "ansible_user_gid": 0, "ansible_user_id": "root", "ansible_user_shell": "/bin/bash", "ansible_user_uid": 0, "ansible_userspace_architecture": "x86_64", "ansible_userspace_bits": "64", "ansible_virtualization_role": "guest", "ansible_virtualization_type": "kvm", "gather_subset": ["all"], "module_setup": true}, "changed": false} META: ran handlers TASK [include_role : storage] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_change_fs.yml:10 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "msg": "", "rc": 0, "results": ["Installed: python3-blivet", "Installed: libblockdev-part-2.19-7.el8.x86_64", "Installed: nss-3.41.0-5.el8.x86_64", "Installed: libblockdev-swap-2.19-7.el8.x86_64", "Installed: userspace-rcu-0.10.1-2.el8.x86_64", "Installed: libblockdev-utils-2.19-7.el8.x86_64", "Installed: mdadm-4.1-4.el8.x86_64", "Installed: nss-softokn-3.41.0-5.el8.x86_64", "Installed: libbytesize-1.4-1.el8.x86_64", "Installed: nss-softokn-freebl-3.41.0-5.el8.x86_64", "Installed: nss-sysinit-3.41.0-5.el8.x86_64", "Installed: nss-util-3.41.0-5.el8.x86_64", "Installed: volume_key-libs-0.3.11-4.el8.x86_64", "Installed: ndctl-62-2.el8.x86_64", "Installed: ndctl-libs-62-2.el8.x86_64", "Installed: blivet-data-1:3.1.0-9.el8.noarch", "Installed: python3-pyparted-1:3.11.0-13.el8.x86_64", "Installed: python3-blivet-1:3.1.0-9.el8.noarch", "Installed: python3-blockdev-2.19-7.el8.x86_64", "Installed: daxctl-libs-62-2.el8.x86_64", "Installed: gdisk-1.0.3-6.el8.x86_64", "Installed: python3-bytesize-1.4-1.el8.x86_64", "Installed: device-mapper-event-8:1.02.155-6.el8.x86_64", "Installed: device-mapper-event-libs-8:1.02.155-6.el8.x86_64", "Installed: libaio-0.3.110-12.el8.x86_64", "Installed: device-mapper-multipath-0.7.8-7.el8.x86_64", "Installed: device-mapper-multipath-libs-0.7.8-7.el8.x86_64", "Installed: device-mapper-persistent-data-0.7.6-1.el8.x86_64", "Installed: lsof-4.91-2.el8.x86_64", "Installed: libblockdev-2.19-7.el8.x86_64", "Installed: lvm2-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-crypto-2.19-7.el8.x86_64", "Installed: libblockdev-dm-2.19-7.el8.x86_64", "Installed: lvm2-libs-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-fs-2.19-7.el8.x86_64", "Installed: libblockdev-kbd-2.19-7.el8.x86_64", "Installed: libblockdev-loop-2.19-7.el8.x86_64", "Installed: libblockdev-lvm-2.19-7.el8.x86_64", "Installed: libblockdev-mdraid-2.19-7.el8.x86_64", "Installed: libblockdev-mpath-2.19-7.el8.x86_64", "Installed: nspr-4.20.0-2.el8.x86_64", "Installed: libblockdev-nvdimm-2.19-7.el8.x86_64"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": [], "changed": false, "leaves": [], "mounts": [], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [], "changed": false, "failed": false, "leaves": [], "mounts": [], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_change_fs.yml:13 included: /tmp/tmpfvho35gc/tests/get_unused_disk.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Find unused disks in the system] ********************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "disks": ["vdb"]} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:8 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"unused_disks": ["vdb"]}, "changed": false} TASK [Exit playbook when there's no unused disks in the system] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:13 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} META: TASK [Print unused disks] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:19 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "unused_disks": [ "vdb" ] } TASK [Create a LVM logical volume with default fs_type] **************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_change_fs.yml:18 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "msg": "Nothing to do", "rc": 0, "results": ["Installed: python3-blivet"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}]}, "ansible_loop_var": "pool", "changed": false, "pool": {"disks": ["vdb"], "name": "foo", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}]}, {'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}]) => {"ansible_facts": {"vol_defaults": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}], "vol_pools": ["foo"], "vols": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}, "ansible_loop_var": "item", "changed": false, "item": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}, {"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}, {'state': 'present', 'type': 'lvm', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid'}]) => {"ansible_facts": {"all_vols": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}]}, "ansible_loop_var": "item", "changed": false, "idx": 0, "item": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"vols_by_pool": {"foo": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}]}}, "ansible_loop_var": "item", "changed": false, "item": {"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}]}]}, "ansible_loop_var": "pool", "changed": false, "idx": 0, "pool": {"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}]}]}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [ { "disks": [ "vdb" ], "name": "foo", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" } ] } ] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": ["create format lvmpv on /dev/vdb", "create device /dev/foo", "create device /dev/mapper/foo-test1", "create format xfs on /dev/mapper/foo-test1"], "changed": true, "leaves": ["/dev/sr0", "/dev/vda1", "/dev/vdc", "/dev/mapper/foo-test1"], "mounts": [{"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted"}], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [ "create format lvmpv on /dev/vdb", "create device /dev/foo", "create device /dev/mapper/foo-test1", "create format xfs on /dev/mapper/foo-test1" ], "changed": true, "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vdc", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted"}, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1"} TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "cmd": ["systemctl", "daemon-reload"], "delta": "0:00:00.617331", "end": "2019-05-26 06:25:06.399788", "rc": 0, "start": "2019-05-26 06:25:05.782457", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_change_fs.yml:30 included: /tmp/tmpfvho35gc/tests/verify-role-results.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:1 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"storage_tests_list": ["verify-volumes"]}, "changed": false} TASK [Print out pool information] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:4 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "_storage_pools_list": [ { "disks": [ "vdb" ], "name": "foo", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" } ] } ] } TASK [Print out volume information] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:9 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} TASK [Ensure the specified devices in the role were correctly configured in the user's system] ************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:14 included: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Read the /etc/fstab file for volume existence] ******************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "cmd": ["cat", "/etc/fstab"], "delta": "0:00:00.010065", "end": "2019-05-26 06:25:10.505616", "rc": 0, "start": "2019-05-26 06:25:10.495551", "stderr": "", "stderr_lines": [], "stdout": "\n#\n# /etc/fstab\n# Created by anaconda on Thu Apr 4 08:39:20 2019\n#\n# Accessible filesystems, by reference, are maintained under '/dev/disk/'.\n# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.\n#\n# After editing this file, run 'systemctl daemon-reload' to update systemd\n# units generated from this file.\n#\nUUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0\n/dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0", "stdout_lines": ["", "#", "# /etc/fstab", "# Created by anaconda on Thu Apr 4 08:39:20 2019", "#", "# Accessible filesystems, by reference, are maintained under '/dev/disk/'.", "# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.", "#", "# After editing this file, run 'systemctl daemon-reload' to update systemd", "# units generated from this file.", "#", "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0", "/dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0"]} TASK [Verify the volumes listed in storage_pools were correctly managed] *********************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:7 included: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"mount_entries": []}, "changed": false} TASK [gather mounts related to each volume in the pool] **************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => {"ansible_facts": {"mount_entries": [[{"block_available": 1290756, "block_size": 4096, "block_total": 1308160, "block_used": 17404, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2621437, "inode_total": 2621440, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 5286936576, "size_total": 5358223360, "uuid": "38b024c6-8112-4e1e-923b-464ba8e44762"}]]}, "ansible_loop_var": "current_volume", "changed": false, "current_volume": {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}} TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:12 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:25 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab.] **************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:38 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" }, "msg": "All assertions passed" } TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:14 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test1", "device": "/dev/mapper/foo-test1"} TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:21 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vda1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vda1", "device": "/dev/vda1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test1", "device": "/dev/mapper/foo-test1"} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:28 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_mount_devs": ["/dev/vda1", "/dev/mapper/foo-test1"], "blivet_mount_devs": ["/dev/mapper/foo-test1"], "removed_volumes": []}, "changed": false} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:33 TASK [Verify the volumes with no pool were correctly managed] ********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:40 TASK [Verify the present volumes with no pool were correctly managed] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:52 TASK [Verify the absent volumes with no pool are not in ansible_mounts] ************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:65 TASK [Change the file system signature on the logical volume created above] ******************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_change_fs.yml:32 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "msg": "Nothing to do", "rc": 0, "results": ["Installed: python3-blivet"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'fs_type': 'xfs', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_type": "xfs", "mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}]}, "ansible_loop_var": "pool", "changed": false, "pool": {"disks": ["vdb"], "name": "foo", "volumes": [{"fs_type": "xfs", "mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'fs_type': 'xfs', 'mount_point': '/opt/test1'}]}, {'name': 'test1', 'size': '5g', 'fs_type': 'xfs', 'mount_point': '/opt/test1'}]) => {"ansible_facts": {"vol_defaults": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}], "vol_pools": ["foo"], "vols": [{"fs_type": "xfs", "mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}, "ansible_loop_var": "item", "changed": false, "item": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_type": "xfs", "mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}, {"fs_type": "xfs", "mount_point": "/opt/test1", "name": "test1", "size": "5g"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'name': 'test1', 'size': '5g', 'fs_type': 'xfs', 'mount_point': '/opt/test1'}, {'state': 'present', 'type': 'lvm', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid'}]) => {"ansible_facts": {"all_vols": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}]}, "ansible_loop_var": "item", "changed": false, "idx": 0, "item": [{"fs_type": "xfs", "mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'fs_type': 'xfs', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"vols_by_pool": {"foo": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}]}}, "ansible_loop_var": "item", "changed": false, "item": {"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_type": "xfs", "mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'fs_type': 'xfs', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}]}]}, "ansible_loop_var": "pool", "changed": false, "idx": 0, "pool": {"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_type": "xfs", "mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}]}]}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [ { "disks": [ "vdb" ], "name": "foo", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" } ] } ] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": [], "changed": false, "leaves": ["/dev/sr0", "/dev/vda1", "/dev/mapper/foo-test1", "/dev/vdc"], "mounts": [{"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted"}], "removed_mounts": [{"device": "/dev/mapper/foo-test1", "path": "/opt/test1"}]} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [], "changed": false, "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/mapper/foo-test1", "/dev/vdc" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "removed_mounts": [ { "device": "/dev/mapper/foo-test1", "path": "/opt/test1" } ] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'path': '/opt/test1', 'device': '/dev/mapper/foo-test1'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "mount_info": {"device": "/dev/mapper/foo-test1", "path": "/opt/test1"}, "name": "/opt/test1", "opts": "defaults", "passno": "0"} TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted"}, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1"} TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "cmd": ["systemctl", "daemon-reload"], "delta": "0:00:00.582817", "end": "2019-05-26 06:25:29.111916", "rc": 0, "start": "2019-05-26 06:25:28.529099", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_change_fs.yml:45 included: /tmp/tmpfvho35gc/tests/verify-role-results.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:1 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"storage_tests_list": ["verify-volumes"]}, "changed": false} TASK [Print out pool information] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:4 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "_storage_pools_list": [ { "disks": [ "vdb" ], "name": "foo", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" } ] } ] } TASK [Print out volume information] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:9 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} TASK [Ensure the specified devices in the role were correctly configured in the user's system] ************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:14 included: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Read the /etc/fstab file for volume existence] ******************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "cmd": ["cat", "/etc/fstab"], "delta": "0:00:00.011739", "end": "2019-05-26 06:25:33.368684", "rc": 0, "start": "2019-05-26 06:25:33.356945", "stderr": "", "stderr_lines": [], "stdout": "\n#\n# /etc/fstab\n# Created by anaconda on Thu Apr 4 08:39:20 2019\n#\n# Accessible filesystems, by reference, are maintained under '/dev/disk/'.\n# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.\n#\n# After editing this file, run 'systemctl daemon-reload' to update systemd\n# units generated from this file.\n#\nUUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0\n/dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0", "stdout_lines": ["", "#", "# /etc/fstab", "# Created by anaconda on Thu Apr 4 08:39:20 2019", "#", "# Accessible filesystems, by reference, are maintained under '/dev/disk/'.", "# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.", "#", "# After editing this file, run 'systemctl daemon-reload' to update systemd", "# units generated from this file.", "#", "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0", "/dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0"]} TASK [Verify the volumes listed in storage_pools were correctly managed] *********************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:7 included: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"mount_entries": []}, "changed": false} TASK [gather mounts related to each volume in the pool] **************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => {"ansible_facts": {"mount_entries": [[{"block_available": 1290756, "block_size": 4096, "block_total": 1308160, "block_used": 17404, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2621437, "inode_total": 2621440, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 5286936576, "size_total": 5358223360, "uuid": "38b024c6-8112-4e1e-923b-464ba8e44762"}]]}, "ansible_loop_var": "current_volume", "changed": false, "current_volume": {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}} TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:12 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:25 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab.] **************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:38 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" }, "msg": "All assertions passed" } TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:14 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test1", "device": "/dev/mapper/foo-test1"} TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:21 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vda1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vda1", "device": "/dev/vda1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test1", "device": "/dev/mapper/foo-test1"} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:28 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_mount_devs": ["/dev/vda1", "/dev/mapper/foo-test1"], "blivet_mount_devs": ["/dev/mapper/foo-test1"], "removed_volumes": []}, "changed": false} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:33 TASK [Verify the volumes with no pool were correctly managed] ********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:40 TASK [Verify the present volumes with no pool were correctly managed] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:52 TASK [Verify the absent volumes with no pool are not in ansible_mounts] ************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:65 META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=86 changed=8 unreachable=0 failed=0 skipped=27 rescued=0 ignored=0 + cd /tmp/tmpfvho35gc/tests; TEST_SUBJECTS=/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TEST_ARTIFACTS=/tmp/linux-system-role-test-work-pull-linux-system-roles_storage-25-db52957-rhel-8-df0ousve/artifacts ansible-playbook -vv --inventory=/usr/share/ansible/inventory/standard-inventory-qcow2 /tmp/tmpfvho35gc/_setup.yml /tmp/tmpfvho35gc/tests/tests_change_fs_use_partitions.yml ansible-playbook 2.8.0 config file = /home/till/.ansible.cfg configured module search path = ['/home/till/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 3.7.3 (default, May 11 2019, 00:38:04) [GCC 9.1.1 20190503 (Red Hat 9.1.1-1)] Using /home/till/.ansible.cfg as config file PLAYBOOK: _setup.yml *************************************************************************************************************************************************************************************** 2 plays in /tmp/tmpfvho35gc/_setup.yml PLAY [Fail when only localhost is available] *************************************************************************************************************************************************************** META: ran handlers TASK [debug] *********************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/_setup.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "groups": { "all": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "localhost": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "subjects": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "ungrouped": [] } } TASK [fail] ************************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:7 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} META: ran handlers META: ran handlers PLAY [Setup repos] ***************************************************************************************************************************************************************************************** META: ran handlers TASK [set up internal repositories] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:16 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true} META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=2 changed=1 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 PLAYBOOK: tests_change_fs_use_partitions.yml *************************************************************************************************************************************************************** 1 plays in /tmp/tmpfvho35gc/tests/tests_change_fs_use_partitions.yml PLAY [all] ************************************************************************************************************************************************************************************************* TASK [Gathering Facts] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/tests_change_fs_use_partitions.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_all_ipv4_addresses": ["10.0.2.15"], "ansible_all_ipv6_addresses": ["fec0::5054:ff:fe12:3456", "fe80::5054:ff:fe12:3456"], "ansible_apparmor": {"status": "disabled"}, "ansible_architecture": "x86_64", "ansible_bios_date": "04/01/2014", "ansible_bios_version": "1.12.0-2.fc30", "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": "ttyS0,115200n8", "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_date_time": {"date": "2019-05-26", "day": "26", "epoch": "1558866440", "hour": "06", "iso8601": "2019-05-26T10:27:20Z", "iso8601_basic": "20190526T062720011484", "iso8601_basic_short": "20190526T062720", "iso8601_micro": "2019-05-26T10:27:20.011801Z", "minute": "27", "month": "05", "second": "20", "time": "06:27:20", "tz": "EDT", "tz_offset": "-0400", "weekday": "Sonntag", "weekday_number": "0", "weeknumber": "20", "year": "2019"}, "ansible_default_ipv4": {"address": "10.0.2.15", "alias": "eth0", "broadcast": "10.0.2.255", "gateway": "10.0.2.2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "netmask": "255.255.255.0", "network": "10.0.2.0", "type": "ether"}, "ansible_default_ipv6": {"address": "fec0::5054:ff:fe12:3456", "gateway": "fe80::2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "prefix": "64", "scope": "site", "type": "ether"}, "ansible_device_links": {"ids": {"sr0": ["ata-QEMU_DVD-ROM_QM00003"]}, "labels": {"sr0": ["cidata"]}, "masters": {}, "uuids": {"sr0": ["2019-05-26-12-25-39-00"], "vda1": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}}, "ansible_devices": {"sr0": {"holders": [], "host": "IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]", "links": {"ids": ["ata-QEMU_DVD-ROM_QM00003"], "labels": ["cidata"], "masters": [], "uuids": ["2019-05-26-12-25-39-00"]}, "model": "QEMU DVD-ROM", "partitions": {}, "removable": "1", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "728", "sectorsize": "2048", "size": "364.00 KB", "support_discard": "0", "vendor": "QEMU", "virtual": 1}, "vda": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {"vda1": {"holders": [], "links": {"ids": [], "labels": [], "masters": [], "uuids": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}, "sectors": "20969439", "sectorsize": 512, "size": "10.00 GB", "start": "2048", "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdb": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdc": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}}, "ansible_distribution": "RedHat", "ansible_distribution_file_parsed": true, "ansible_distribution_file_path": "/etc/redhat-release", "ansible_distribution_file_search_string": "Red Hat", "ansible_distribution_file_variety": "RedHat", "ansible_distribution_major_version": "8", "ansible_distribution_release": "Ootpa", "ansible_distribution_version": "8.0", "ansible_dns": {"nameservers": ["10.0.2.3"], "search": ["virt.pnr.lab.eng.rdu2.redhat.com"]}, "ansible_domain": "virt.pnr.lab.eng.rdu2.redhat.com", "ansible_effective_group_id": 0, "ansible_effective_user_id": 0, "ansible_env": {"DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "HOME": "/root", "LANG": "de_DE.UTF-8", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "LOGNAME": "root", "LS_COLORS": "rs=0:di=38;5;33:ln=38;5;51:mh=00:pi=40;38;5;11:so=38;5;13:do=38;5;5:bd=48;5;232;38;5;11:cd=48;5;232;38;5;3:or=48;5;232;38;5;9:mi=01;05;37;41:su=48;5;196;38;5;15:sg=48;5;11;38;5;16:ca=48;5;196;38;5;226:tw=48;5;10;38;5;16:ow=48;5;10;38;5;21:st=48;5;21;38;5;15:ex=38;5;40:*.tar=38;5;9:*.tgz=38;5;9:*.arc=38;5;9:*.arj=38;5;9:*.taz=38;5;9:*.lha=38;5;9:*.lz4=38;5;9:*.lzh=38;5;9:*.lzma=38;5;9:*.tlz=38;5;9:*.txz=38;5;9:*.tzo=38;5;9:*.t7z=38;5;9:*.zip=38;5;9:*.z=38;5;9:*.dz=38;5;9:*.gz=38;5;9:*.lrz=38;5;9:*.lz=38;5;9:*.lzo=38;5;9:*.xz=38;5;9:*.zst=38;5;9:*.tzst=38;5;9:*.bz2=38;5;9:*.bz=38;5;9:*.tbz=38;5;9:*.tbz2=38;5;9:*.tz=38;5;9:*.deb=38;5;9:*.rpm=38;5;9:*.jar=38;5;9:*.war=38;5;9:*.ear=38;5;9:*.sar=38;5;9:*.rar=38;5;9:*.alz=38;5;9:*.ace=38;5;9:*.zoo=38;5;9:*.cpio=38;5;9:*.7z=38;5;9:*.rz=38;5;9:*.cab=38;5;9:*.wim=38;5;9:*.swm=38;5;9:*.dwm=38;5;9:*.esd=38;5;9:*.jpg=38;5;13:*.jpeg=38;5;13:*.mjpg=38;5;13:*.mjpeg=38;5;13:*.gif=38;5;13:*.bmp=38;5;13:*.pbm=38;5;13:*.pgm=38;5;13:*.ppm=38;5;13:*.tga=38;5;13:*.xbm=38;5;13:*.xpm=38;5;13:*.tif=38;5;13:*.tiff=38;5;13:*.png=38;5;13:*.svg=38;5;13:*.svgz=38;5;13:*.mng=38;5;13:*.pcx=38;5;13:*.mov=38;5;13:*.mpg=38;5;13:*.mpeg=38;5;13:*.m2v=38;5;13:*.mkv=38;5;13:*.webm=38;5;13:*.ogm=38;5;13:*.mp4=38;5;13:*.m4v=38;5;13:*.mp4v=38;5;13:*.vob=38;5;13:*.qt=38;5;13:*.nuv=38;5;13:*.wmv=38;5;13:*.asf=38;5;13:*.rm=38;5;13:*.rmvb=38;5;13:*.flc=38;5;13:*.avi=38;5;13:*.fli=38;5;13:*.flv=38;5;13:*.gl=38;5;13:*.dl=38;5;13:*.xcf=38;5;13:*.xwd=38;5;13:*.yuv=38;5;13:*.cgm=38;5;13:*.emf=38;5;13:*.ogv=38;5;13:*.ogx=38;5;13:*.aac=38;5;45:*.au=38;5;45:*.flac=38;5;45:*.m4a=38;5;45:*.mid=38;5;45:*.midi=38;5;45:*.mka=38;5;45:*.mp3=38;5;45:*.mpc=38;5;45:*.ogg=38;5;45:*.ra=38;5;45:*.wav=38;5;45:*.oga=38;5;45:*.opus=38;5;45:*.spx=38;5;45:*.xspf=38;5;45:", "MAIL": "/var/mail/root", "PATH": "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "PWD": "/root", "SELINUX_LEVEL_REQUESTED": "", "SELINUX_ROLE_REQUESTED": "", "SELINUX_USE_CURRENT_RANGE": "", "SHELL": "/bin/bash", "SHLVL": "2", "SSH_CLIENT": "10.0.2.2 46152 22", "SSH_CONNECTION": "10.0.2.2 46152 10.0.2.15 22", "SSH_TTY": "/dev/pts/0", "TERM": "xterm-256color", "USER": "root", "XDG_RUNTIME_DIR": "/run/user/0", "XDG_SESSION_ID": "1", "_": "/usr/libexec/platform-python"}, "ansible_eth0": {"active": true, "device": "eth0", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "off [requested on]", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "off [fixed]", "netns_local": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "off [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "on [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "off", "tcp_segmentation_offload": "off", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "tx_checksumming": "off", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "off [fixed]", "tx_nocache_copy": "off", "tx_scatter_gather": "off [fixed]", "tx_scatter_gather_fraglist": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_tcp6_segmentation": "off [fixed]", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off [fixed]", "tx_tcp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "off [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "10.0.2.15", "broadcast": "10.0.2.255", "netmask": "255.255.255.0", "network": "10.0.2.0"}, "ipv6": [{"address": "fec0::5054:ff:fe12:3456", "prefix": "64", "scope": "site"}, {"address": "fe80::5054:ff:fe12:3456", "prefix": "64", "scope": "link"}], "macaddress": "52:54:00:12:34:56", "module": "virtio_net", "mtu": 1500, "pciid": "virtio0", "promisc": false, "speed": -1, "timestamping": ["tx_software", "rx_software", "software"], "type": "ether"}, "ansible_fibre_channel_wwn": [], "ansible_fips": false, "ansible_form_factor": "Other", "ansible_fqdn": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_hostname": "ibm-p8-kvm-03-guest-02", "ansible_hostnqn": "", "ansible_interfaces": ["eth0", "lo"], "ansible_is_chroot": false, "ansible_iscsi_iqn": "", "ansible_kernel": "4.18.0-80.el8.x86_64", "ansible_lo": {"active": true, "device": "lo", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "on", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "on [fixed]", "netns_local": "on [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "on [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "off [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "on", "tcp_segmentation_offload": "on", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "tx_checksumming": "on", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "on [fixed]", "tx_nocache_copy": "off [fixed]", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tx_sctp_segmentation": "on", "tx_tcp6_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp_segmentation": "on", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "on [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "127.0.0.1", "broadcast": "host", "netmask": "255.0.0.0", "network": "127.0.0.0"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "mtu": 65536, "promisc": false, "timestamping": ["tx_software", "rx_software", "software"], "type": "loopback"}, "ansible_local": {}, "ansible_lsb": {}, "ansible_machine": "x86_64", "ansible_machine_id": "40775d7f510b403895226b7e1af4ac3e", "ansible_memfree_mb": 392, "ansible_memory_mb": {"nocache": {"free": 624, "used": 196}, "real": {"free": 392, "total": 820, "used": 428}, "swap": {"cached": 0, "free": 0, "total": 0, "used": 0}}, "ansible_memtotal_mb": 820, "ansible_mounts": [{"block_available": 2291651, "block_size": 4096, "block_total": 2618619, "block_used": 326968, "device": "/dev/vda1", "fstype": "xfs", "inode_available": 5207721, "inode_total": 5242352, "inode_used": 34631, "mount": "/", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 9386602496, "size_total": 10725863424, "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}], "ansible_nodename": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_os_family": "RedHat", "ansible_pkg_mgr": "dnf", "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": ["ttyS0", "ttyS0,115200n8"], "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_processor": ["0", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "1", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "2", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz"], "ansible_processor_cores": 1, "ansible_processor_count": 3, "ansible_processor_threads_per_core": 1, "ansible_processor_vcpus": 3, "ansible_product_name": "Standard PC (i440FX + PIIX, 1996)", "ansible_product_serial": "NA", "ansible_product_uuid": "NA", "ansible_product_version": "pc-i440fx-3.1", "ansible_python": {"executable": "/usr/libexec/platform-python", "has_sslcontext": true, "type": "cpython", "version": {"major": 3, "micro": 8, "minor": 6, "releaselevel": "final", "serial": 0}, "version_info": [3, 6, 8, "final", 0]}, "ansible_python_version": "3.6.8", "ansible_real_group_id": 0, "ansible_real_user_id": 0, "ansible_selinux": {"config_mode": "enforcing", "mode": "enforcing", "policyvers": 31, "status": "enabled", "type": "targeted"}, "ansible_selinux_python_present": true, "ansible_service_mgr": "systemd", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBGj/oL5U44n8hpEsyLJ6KDCnmAHAgzTvgzz6pU6+kQm8YWJKQS/+td5rj7inYhp5MuX7AD7vLDFcG7+bSwjoYv0=", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIBTPx6MLKGwBr5S+y3lIuGJhCXQYiKdcAELIt/kiBmmf", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABAQDQ0JV2rEzXvIfKoCzKxDThiBTj92b9BKN7E6nMULF96bx9XhDoLQQ0Rkzt6GEWhbHi1RK8B3rICFvi6RcAokKICO0u5mBCEu4SDwtmh5DehJxLNC+I4mKryHwpPZ/KGK5vj/JFASOpZHpy6qxXBRqwYwnIPc1FjnTXtQcIhN4WbVXj7OWJNG62cw+QOm6wkL9kLGdnCyd2IQIO6By4GotnxIfnyAEHGnm0+ZNBqYX/7Jwjp45FX8dU0S+TOdZ2XV9NjQI88wA9proYsvednPE+ty+ShcUXNuUDoiEf1kp4Jy2lZXnhIT19kOs9ozOY30CHtVYvNoU8Qz8gwng4zQu9", "ansible_swapfree_mb": 0, "ansible_swaptotal_mb": 0, "ansible_system": "Linux", "ansible_system_capabilities": ["cap_chown", "cap_dac_override", "cap_dac_read_search", "cap_fowner", "cap_fsetid", "cap_kill", "cap_setgid", "cap_setuid", "cap_setpcap", "cap_linux_immutable", "cap_net_bind_service", "cap_net_broadcast", "cap_net_admin", "cap_net_raw", "cap_ipc_lock", "cap_ipc_owner", "cap_sys_module", "cap_sys_rawio", "cap_sys_chroot", "cap_sys_ptrace", "cap_sys_pacct", "cap_sys_admin", "cap_sys_boot", "cap_sys_nice", "cap_sys_resource", "cap_sys_time", "cap_sys_tty_config", "cap_mknod", "cap_lease", "cap_audit_write", "cap_audit_control", "cap_setfcap", "cap_mac_override", "cap_mac_admin", "cap_syslog", "cap_wake_alarm", "cap_block_suspend", "cap_audit_read+ep"], "ansible_system_capabilities_enforced": "True", "ansible_system_vendor": "QEMU", "ansible_uptime_seconds": 95, "ansible_user_dir": "/root", "ansible_user_gecos": "root", "ansible_user_gid": 0, "ansible_user_id": "root", "ansible_user_shell": "/bin/bash", "ansible_user_uid": 0, "ansible_userspace_architecture": "x86_64", "ansible_userspace_bits": "64", "ansible_virtualization_role": "guest", "ansible_virtualization_type": "kvm", "gather_subset": ["all"], "module_setup": true}, "changed": false} META: ran handlers TASK [include_role : storage] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_change_fs_use_partitions.yml:11 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "msg": "", "rc": 0, "results": ["Installed: python3-blivet", "Installed: libblockdev-part-2.19-7.el8.x86_64", "Installed: nss-3.41.0-5.el8.x86_64", "Installed: libblockdev-swap-2.19-7.el8.x86_64", "Installed: userspace-rcu-0.10.1-2.el8.x86_64", "Installed: libblockdev-utils-2.19-7.el8.x86_64", "Installed: mdadm-4.1-4.el8.x86_64", "Installed: nss-softokn-3.41.0-5.el8.x86_64", "Installed: libbytesize-1.4-1.el8.x86_64", "Installed: nss-softokn-freebl-3.41.0-5.el8.x86_64", "Installed: nss-sysinit-3.41.0-5.el8.x86_64", "Installed: nss-util-3.41.0-5.el8.x86_64", "Installed: volume_key-libs-0.3.11-4.el8.x86_64", "Installed: ndctl-62-2.el8.x86_64", "Installed: ndctl-libs-62-2.el8.x86_64", "Installed: blivet-data-1:3.1.0-9.el8.noarch", "Installed: python3-pyparted-1:3.11.0-13.el8.x86_64", "Installed: python3-blivet-1:3.1.0-9.el8.noarch", "Installed: python3-blockdev-2.19-7.el8.x86_64", "Installed: daxctl-libs-62-2.el8.x86_64", "Installed: gdisk-1.0.3-6.el8.x86_64", "Installed: python3-bytesize-1.4-1.el8.x86_64", "Installed: device-mapper-event-8:1.02.155-6.el8.x86_64", "Installed: device-mapper-event-libs-8:1.02.155-6.el8.x86_64", "Installed: libaio-0.3.110-12.el8.x86_64", "Installed: device-mapper-multipath-0.7.8-7.el8.x86_64", "Installed: device-mapper-multipath-libs-0.7.8-7.el8.x86_64", "Installed: device-mapper-persistent-data-0.7.6-1.el8.x86_64", "Installed: lsof-4.91-2.el8.x86_64", "Installed: libblockdev-2.19-7.el8.x86_64", "Installed: lvm2-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-crypto-2.19-7.el8.x86_64", "Installed: libblockdev-dm-2.19-7.el8.x86_64", "Installed: lvm2-libs-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-fs-2.19-7.el8.x86_64", "Installed: libblockdev-kbd-2.19-7.el8.x86_64", "Installed: libblockdev-loop-2.19-7.el8.x86_64", "Installed: libblockdev-lvm-2.19-7.el8.x86_64", "Installed: libblockdev-mdraid-2.19-7.el8.x86_64", "Installed: libblockdev-mpath-2.19-7.el8.x86_64", "Installed: nspr-4.20.0-2.el8.x86_64", "Installed: libblockdev-nvdimm-2.19-7.el8.x86_64"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": [], "changed": false, "leaves": [], "mounts": [], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [], "changed": false, "failed": false, "leaves": [], "mounts": [], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_change_fs_use_partitions.yml:14 included: /tmp/tmpfvho35gc/tests/get_unused_disk.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Find unused disks in the system] ********************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "disks": ["vdb"]} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:8 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"unused_disks": ["vdb"]}, "changed": false} TASK [Exit playbook when there's no unused disks in the system] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:13 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} META: TASK [Print unused disks] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:19 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "unused_disks": [ "vdb" ] } TASK [Create an LVM partition with the default file system type] ******************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/tests_change_fs_use_partitions.yml:19 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "msg": "Nothing to do", "rc": 0, "results": ["Installed: python3-blivet"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'name': 'bar', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "bar", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}]}, "ansible_loop_var": "pool", "changed": false, "pool": {"disks": ["vdb"], "name": "bar", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'state': 'present', 'type': 'lvm', 'name': 'bar', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}]}, {'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}]) => {"ansible_facts": {"vol_defaults": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}], "vol_pools": ["bar"], "vols": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}, "ansible_loop_var": "item", "changed": false, "item": [{"disks": ["vdb"], "name": "bar", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}, {"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}, {'state': 'present', 'type': 'lvm', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid'}]) => {"ansible_facts": {"all_vols": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm"}]}, "ansible_loop_var": "item", "changed": false, "idx": 0, "item": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'bar', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"vols_by_pool": {"bar": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm"}]}}, "ansible_loop_var": "item", "changed": false, "item": {"disks": ["vdb"], "name": "bar", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'bar', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "bar", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm"}]}]}, "ansible_loop_var": "pool", "changed": false, "idx": 0, "pool": {"disks": ["vdb"], "name": "bar", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": [{"disks": ["vdb"], "name": "bar", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm"}]}]}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [ { "disks": [ "vdb" ], "name": "bar", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm" } ] } ] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": ["create format lvmpv on /dev/vdb", "create device /dev/bar", "create device /dev/mapper/bar-test1", "create format xfs on /dev/mapper/bar-test1"], "changed": true, "leaves": ["/dev/sr0", "/dev/vda1", "/dev/vdc", "/dev/mapper/bar-test1"], "mounts": [{"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/bar-test1", "state": "mounted"}], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [ "create format lvmpv on /dev/vdb", "create device /dev/bar", "create device /dev/mapper/bar-test1", "create format xfs on /dev/mapper/bar-test1" ], "changed": true, "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vdc", "/dev/mapper/bar-test1" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/bar-test1", "state": "mounted" } ], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'src': '/dev/mapper/bar-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/bar-test1", "state": "mounted"}, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/bar-test1"} TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "cmd": ["systemctl", "daemon-reload"], "delta": "0:00:00.653317", "end": "2019-05-26 06:28:31.343437", "rc": 0, "start": "2019-05-26 06:28:30.690120", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_change_fs_use_partitions.yml:31 included: /tmp/tmpfvho35gc/tests/verify-role-results.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:1 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"storage_tests_list": ["verify-volumes"]}, "changed": false} TASK [Print out pool information] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:4 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "_storage_pools_list": [ { "disks": [ "vdb" ], "name": "bar", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm" } ] } ] } TASK [Print out volume information] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:9 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} TASK [Ensure the specified devices in the role were correctly configured in the user's system] ************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:14 included: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Read the /etc/fstab file for volume existence] ******************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "cmd": ["cat", "/etc/fstab"], "delta": "0:00:00.014124", "end": "2019-05-26 06:28:35.852730", "rc": 0, "start": "2019-05-26 06:28:35.838606", "stderr": "", "stderr_lines": [], "stdout": "\n#\n# /etc/fstab\n# Created by anaconda on Thu Apr 4 08:39:20 2019\n#\n# Accessible filesystems, by reference, are maintained under '/dev/disk/'.\n# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.\n#\n# After editing this file, run 'systemctl daemon-reload' to update systemd\n# units generated from this file.\n#\nUUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0\n/dev/mapper/bar-test1 /opt/test1 xfs defaults 0 0", "stdout_lines": ["", "#", "# /etc/fstab", "# Created by anaconda on Thu Apr 4 08:39:20 2019", "#", "# Accessible filesystems, by reference, are maintained under '/dev/disk/'.", "# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.", "#", "# After editing this file, run 'systemctl daemon-reload' to update systemd", "# units generated from this file.", "#", "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0", "/dev/mapper/bar-test1 /opt/test1 xfs defaults 0 0"]} TASK [Verify the volumes listed in storage_pools were correctly managed] *********************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:7 included: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"mount_entries": []}, "changed": false} TASK [gather mounts related to each volume in the pool] **************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'bar'}) => {"ansible_facts": {"mount_entries": [[{"block_available": 1290756, "block_size": 4096, "block_total": 1308160, "block_used": 17404, "device": "/dev/mapper/bar-test1", "fstype": "xfs", "inode_available": 2621437, "inode_total": 2621440, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 5286936576, "size_total": 5358223360, "uuid": "b83b2d4b-77dc-439b-9a20-37660e68c406"}]]}, "ansible_loop_var": "current_volume", "changed": false, "current_volume": {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm"}} TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:12 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'bar'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:25 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'bar'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab.] **************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:38 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'bar'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm" }, "msg": "All assertions passed" } TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:14 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/bar-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/bar-test1", "device": "/dev/mapper/bar-test1"} TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:21 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vda1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vda1", "device": "/dev/vda1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/bar-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/bar-test1", "device": "/dev/mapper/bar-test1"} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:28 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_mount_devs": ["/dev/vda1", "/dev/mapper/bar-test1"], "blivet_mount_devs": ["/dev/mapper/bar-test1"], "removed_volumes": []}, "changed": false} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:33 TASK [Verify the volumes with no pool were correctly managed] ********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:40 TASK [Verify the present volumes with no pool were correctly managed] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:52 TASK [Verify the absent volumes with no pool are not in ansible_mounts] ************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:65 TASK [Change the LVM partition file system type to "ext4"] ************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/tests_change_fs_use_partitions.yml:33 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "msg": "Nothing to do", "rc": 0, "results": ["Installed: python3-blivet"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'name': 'bar', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'fs_type': 'ext4', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "bar", "state": "present", "type": "lvm", "volumes": [{"fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}]}, "ansible_loop_var": "pool", "changed": false, "pool": {"disks": ["vdb"], "name": "bar", "volumes": [{"fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'state': 'present', 'type': 'lvm', 'name': 'bar', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'fs_type': 'ext4', 'mount_point': '/opt/test1'}]}, {'name': 'test1', 'size': '5g', 'fs_type': 'ext4', 'mount_point': '/opt/test1'}]) => {"ansible_facts": {"vol_defaults": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}], "vol_pools": ["bar"], "vols": [{"fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}, "ansible_loop_var": "item", "changed": false, "item": [{"disks": ["vdb"], "name": "bar", "state": "present", "type": "lvm", "volumes": [{"fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}, {"fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "5g"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'name': 'test1', 'size': '5g', 'fs_type': 'ext4', 'mount_point': '/opt/test1'}, {'state': 'present', 'type': 'lvm', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid'}]) => {"ansible_facts": {"all_vols": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm"}]}, "ansible_loop_var": "item", "changed": false, "idx": 0, "item": [{"fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'bar', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'fs_type': 'ext4', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"vols_by_pool": {"bar": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm"}]}}, "ansible_loop_var": "item", "changed": false, "item": {"disks": ["vdb"], "name": "bar", "state": "present", "type": "lvm", "volumes": [{"fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'bar', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'fs_type': 'ext4', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "bar", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm"}]}]}, "ansible_loop_var": "pool", "changed": false, "idx": 0, "pool": {"disks": ["vdb"], "name": "bar", "state": "present", "type": "lvm", "volumes": [{"fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": [{"disks": ["vdb"], "name": "bar", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm"}]}]}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [ { "disks": [ "vdb" ], "name": "bar", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm" } ] } ] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": ["destroy format xfs on /dev/mapper/bar-test1", "create format ext4 on /dev/mapper/bar-test1"], "changed": true, "leaves": ["/dev/sr0", "/dev/vda1", "/dev/mapper/bar-test1", "/dev/vdc"], "mounts": [{"dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/bar-test1", "state": "mounted"}], "removed_mounts": [{"device": "/dev/mapper/bar-test1", "path": "/opt/test1"}, {"device": "/dev/mapper/bar-test1", "path": "/opt/test1"}]} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [ "destroy format xfs on /dev/mapper/bar-test1", "create format ext4 on /dev/mapper/bar-test1" ], "changed": true, "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/mapper/bar-test1", "/dev/vdc" ], "mounts": [ { "dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/bar-test1", "state": "mounted" } ], "removed_mounts": [ { "device": "/dev/mapper/bar-test1", "path": "/opt/test1" }, { "device": "/dev/mapper/bar-test1", "path": "/opt/test1" } ] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'path': '/opt/test1', 'device': '/dev/mapper/bar-test1'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "mount_info": {"device": "/dev/mapper/bar-test1", "path": "/opt/test1"}, "name": "/opt/test1", "opts": "defaults", "passno": "0"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'path': '/opt/test1', 'device': '/dev/mapper/bar-test1'}) => {"ansible_loop_var": "mount_info", "changed": false, "dump": "0", "fstab": "/etc/fstab", "mount_info": {"device": "/dev/mapper/bar-test1", "path": "/opt/test1"}, "name": "/opt/test1", "opts": "defaults", "passno": "0"} TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'src': '/dev/mapper/bar-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": {"dump": 0, "fstype": "ext4", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/bar-test1", "state": "mounted"}, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/bar-test1"} TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "cmd": ["systemctl", "daemon-reload"], "delta": "0:00:00.723944", "end": "2019-05-26 06:28:59.890527", "rc": 0, "start": "2019-05-26 06:28:59.166583", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_change_fs_use_partitions.yml:46 included: /tmp/tmpfvho35gc/tests/verify-role-results.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:1 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"storage_tests_list": ["verify-volumes"]}, "changed": false} TASK [Print out pool information] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:4 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "_storage_pools_list": [ { "disks": [ "vdb" ], "name": "bar", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm" } ] } ] } TASK [Print out volume information] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:9 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} TASK [Ensure the specified devices in the role were correctly configured in the user's system] ************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:14 included: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Read the /etc/fstab file for volume existence] ******************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "cmd": ["cat", "/etc/fstab"], "delta": "0:00:00.009664", "end": "2019-05-26 06:29:04.617799", "rc": 0, "start": "2019-05-26 06:29:04.608135", "stderr": "", "stderr_lines": [], "stdout": "\n#\n# /etc/fstab\n# Created by anaconda on Thu Apr 4 08:39:20 2019\n#\n# Accessible filesystems, by reference, are maintained under '/dev/disk/'.\n# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.\n#\n# After editing this file, run 'systemctl daemon-reload' to update systemd\n# units generated from this file.\n#\nUUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0\n/dev/mapper/bar-test1 /opt/test1 ext4 defaults 0 0", "stdout_lines": ["", "#", "# /etc/fstab", "# Created by anaconda on Thu Apr 4 08:39:20 2019", "#", "# Accessible filesystems, by reference, are maintained under '/dev/disk/'.", "# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.", "#", "# After editing this file, run 'systemctl daemon-reload' to update systemd", "# units generated from this file.", "#", "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0", "/dev/mapper/bar-test1 /opt/test1 ext4 defaults 0 0"]} TASK [Verify the volumes listed in storage_pools were correctly managed] *********************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:7 included: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"mount_entries": []}, "changed": false} TASK [gather mounts related to each volume in the pool] **************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'ext4', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'bar'}) => {"ansible_facts": {"mount_entries": [[{"block_available": 1199010, "block_size": 4096, "block_total": 1273760, "block_used": 74750, "device": "/dev/mapper/bar-test1", "fstype": "ext4", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime", "size_available": 4911144960, "size_total": 5217320960, "uuid": "017aa9c3-f6fd-46cf-844c-e199ffdc5633"}]]}, "ansible_loop_var": "current_volume", "changed": false, "current_volume": {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm"}} TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:12 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'ext4', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'bar'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:25 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'ext4', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'bar'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab.] **************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:38 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'ext4', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'bar'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "bar", "size": "5g", "state": "present", "type": "lvm" }, "msg": "All assertions passed" } TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:14 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/bar-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/bar-test1", "device": "/dev/mapper/bar-test1"} TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:21 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vda1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vda1", "device": "/dev/vda1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/bar-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/bar-test1", "device": "/dev/mapper/bar-test1"} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:28 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_mount_devs": ["/dev/vda1", "/dev/mapper/bar-test1"], "blivet_mount_devs": ["/dev/mapper/bar-test1"], "removed_volumes": []}, "changed": false} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:33 TASK [Verify the volumes with no pool were correctly managed] ********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:40 TASK [Verify the present volumes with no pool were correctly managed] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:52 TASK [Verify the absent volumes with no pool are not in ansible_mounts] ************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:65 META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=86 changed=9 unreachable=0 failed=0 skipped=27 rescued=0 ignored=0 + cd /tmp/tmpfvho35gc/tests; TEST_SUBJECTS=/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TEST_ARTIFACTS=/tmp/linux-system-role-test-work-pull-linux-system-roles_storage-25-db52957-rhel-8-df0ousve/artifacts ansible-playbook -vv --inventory=/usr/share/ansible/inventory/standard-inventory-qcow2 /tmp/tmpfvho35gc/_setup.yml /tmp/tmpfvho35gc/tests/tests_change_mount.yml ansible-playbook 2.8.0 config file = /home/till/.ansible.cfg configured module search path = ['/home/till/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 3.7.3 (default, May 11 2019, 00:38:04) [GCC 9.1.1 20190503 (Red Hat 9.1.1-1)] Using /home/till/.ansible.cfg as config file PLAYBOOK: _setup.yml *************************************************************************************************************************************************************************************** 2 plays in /tmp/tmpfvho35gc/_setup.yml PLAY [Fail when only localhost is available] *************************************************************************************************************************************************************** META: ran handlers TASK [debug] *********************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/_setup.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "groups": { "all": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "localhost": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "subjects": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "ungrouped": [] } } TASK [fail] ************************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:7 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} META: ran handlers META: ran handlers PLAY [Setup repos] ***************************************************************************************************************************************************************************************** META: ran handlers TASK [set up internal repositories] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:16 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true} META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=2 changed=1 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 PLAYBOOK: tests_change_mount.yml *************************************************************************************************************************************************************************** 1 plays in /tmp/tmpfvho35gc/tests/tests_change_mount.yml PLAY [all] ************************************************************************************************************************************************************************************************* TASK [Gathering Facts] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/tests_change_mount.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_all_ipv4_addresses": ["10.0.2.15"], "ansible_all_ipv6_addresses": ["fec0::5054:ff:fe12:3456", "fe80::5054:ff:fe12:3456"], "ansible_apparmor": {"status": "disabled"}, "ansible_architecture": "x86_64", "ansible_bios_date": "04/01/2014", "ansible_bios_version": "1.12.0-2.fc30", "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": "ttyS0,115200n8", "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_date_time": {"date": "2019-05-26", "day": "26", "epoch": "1558866652", "hour": "06", "iso8601": "2019-05-26T10:30:52Z", "iso8601_basic": "20190526T063052669639", "iso8601_basic_short": "20190526T063052", "iso8601_micro": "2019-05-26T10:30:52.669880Z", "minute": "30", "month": "05", "second": "52", "time": "06:30:52", "tz": "EDT", "tz_offset": "-0400", "weekday": "Sonntag", "weekday_number": "0", "weeknumber": "20", "year": "2019"}, "ansible_default_ipv4": {"address": "10.0.2.15", "alias": "eth0", "broadcast": "10.0.2.255", "gateway": "10.0.2.2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "netmask": "255.255.255.0", "network": "10.0.2.0", "type": "ether"}, "ansible_default_ipv6": {"address": "fec0::5054:ff:fe12:3456", "gateway": "fe80::2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "prefix": "64", "scope": "site", "type": "ether"}, "ansible_device_links": {"ids": {"sr0": ["ata-QEMU_DVD-ROM_QM00003"]}, "labels": {"sr0": ["cidata"]}, "masters": {}, "uuids": {"sr0": ["2019-05-26-12-29-10-00"], "vda1": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}}, "ansible_devices": {"sr0": {"holders": [], "host": "IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]", "links": {"ids": ["ata-QEMU_DVD-ROM_QM00003"], "labels": ["cidata"], "masters": [], "uuids": ["2019-05-26-12-29-10-00"]}, "model": "QEMU DVD-ROM", "partitions": {}, "removable": "1", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "728", "sectorsize": "2048", "size": "364.00 KB", "support_discard": "0", "vendor": "QEMU", "virtual": 1}, "vda": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {"vda1": {"holders": [], "links": {"ids": [], "labels": [], "masters": [], "uuids": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}, "sectors": "20969439", "sectorsize": 512, "size": "10.00 GB", "start": "2048", "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdb": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdc": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}}, "ansible_distribution": "RedHat", "ansible_distribution_file_parsed": true, "ansible_distribution_file_path": "/etc/redhat-release", "ansible_distribution_file_search_string": "Red Hat", "ansible_distribution_file_variety": "RedHat", "ansible_distribution_major_version": "8", "ansible_distribution_release": "Ootpa", "ansible_distribution_version": "8.0", "ansible_dns": {"nameservers": ["10.0.2.3"], "search": ["virt.pnr.lab.eng.rdu2.redhat.com"]}, "ansible_domain": "virt.pnr.lab.eng.rdu2.redhat.com", "ansible_effective_group_id": 0, "ansible_effective_user_id": 0, "ansible_env": {"DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "HOME": "/root", "LANG": "de_DE.UTF-8", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "LOGNAME": "root", "LS_COLORS": "rs=0:di=38;5;33:ln=38;5;51:mh=00:pi=40;38;5;11:so=38;5;13:do=38;5;5:bd=48;5;232;38;5;11:cd=48;5;232;38;5;3:or=48;5;232;38;5;9:mi=01;05;37;41:su=48;5;196;38;5;15:sg=48;5;11;38;5;16:ca=48;5;196;38;5;226:tw=48;5;10;38;5;16:ow=48;5;10;38;5;21:st=48;5;21;38;5;15:ex=38;5;40:*.tar=38;5;9:*.tgz=38;5;9:*.arc=38;5;9:*.arj=38;5;9:*.taz=38;5;9:*.lha=38;5;9:*.lz4=38;5;9:*.lzh=38;5;9:*.lzma=38;5;9:*.tlz=38;5;9:*.txz=38;5;9:*.tzo=38;5;9:*.t7z=38;5;9:*.zip=38;5;9:*.z=38;5;9:*.dz=38;5;9:*.gz=38;5;9:*.lrz=38;5;9:*.lz=38;5;9:*.lzo=38;5;9:*.xz=38;5;9:*.zst=38;5;9:*.tzst=38;5;9:*.bz2=38;5;9:*.bz=38;5;9:*.tbz=38;5;9:*.tbz2=38;5;9:*.tz=38;5;9:*.deb=38;5;9:*.rpm=38;5;9:*.jar=38;5;9:*.war=38;5;9:*.ear=38;5;9:*.sar=38;5;9:*.rar=38;5;9:*.alz=38;5;9:*.ace=38;5;9:*.zoo=38;5;9:*.cpio=38;5;9:*.7z=38;5;9:*.rz=38;5;9:*.cab=38;5;9:*.wim=38;5;9:*.swm=38;5;9:*.dwm=38;5;9:*.esd=38;5;9:*.jpg=38;5;13:*.jpeg=38;5;13:*.mjpg=38;5;13:*.mjpeg=38;5;13:*.gif=38;5;13:*.bmp=38;5;13:*.pbm=38;5;13:*.pgm=38;5;13:*.ppm=38;5;13:*.tga=38;5;13:*.xbm=38;5;13:*.xpm=38;5;13:*.tif=38;5;13:*.tiff=38;5;13:*.png=38;5;13:*.svg=38;5;13:*.svgz=38;5;13:*.mng=38;5;13:*.pcx=38;5;13:*.mov=38;5;13:*.mpg=38;5;13:*.mpeg=38;5;13:*.m2v=38;5;13:*.mkv=38;5;13:*.webm=38;5;13:*.ogm=38;5;13:*.mp4=38;5;13:*.m4v=38;5;13:*.mp4v=38;5;13:*.vob=38;5;13:*.qt=38;5;13:*.nuv=38;5;13:*.wmv=38;5;13:*.asf=38;5;13:*.rm=38;5;13:*.rmvb=38;5;13:*.flc=38;5;13:*.avi=38;5;13:*.fli=38;5;13:*.flv=38;5;13:*.gl=38;5;13:*.dl=38;5;13:*.xcf=38;5;13:*.xwd=38;5;13:*.yuv=38;5;13:*.cgm=38;5;13:*.emf=38;5;13:*.ogv=38;5;13:*.ogx=38;5;13:*.aac=38;5;45:*.au=38;5;45:*.flac=38;5;45:*.m4a=38;5;45:*.mid=38;5;45:*.midi=38;5;45:*.mka=38;5;45:*.mp3=38;5;45:*.mpc=38;5;45:*.ogg=38;5;45:*.ra=38;5;45:*.wav=38;5;45:*.oga=38;5;45:*.opus=38;5;45:*.spx=38;5;45:*.xspf=38;5;45:", "MAIL": "/var/mail/root", "PATH": "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "PWD": "/root", "SELINUX_LEVEL_REQUESTED": "", "SELINUX_ROLE_REQUESTED": "", "SELINUX_USE_CURRENT_RANGE": "", "SHELL": "/bin/bash", "SHLVL": "2", "SSH_CLIENT": "10.0.2.2 42848 22", "SSH_CONNECTION": "10.0.2.2 42848 10.0.2.15 22", "SSH_TTY": "/dev/pts/0", "TERM": "xterm-256color", "USER": "root", "XDG_RUNTIME_DIR": "/run/user/0", "XDG_SESSION_ID": "1", "_": "/usr/libexec/platform-python"}, "ansible_eth0": {"active": true, "device": "eth0", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "off [requested on]", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "off [fixed]", "netns_local": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "off [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "on [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "off", "tcp_segmentation_offload": "off", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "tx_checksumming": "off", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "off [fixed]", "tx_nocache_copy": "off", "tx_scatter_gather": "off [fixed]", "tx_scatter_gather_fraglist": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_tcp6_segmentation": "off [fixed]", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off [fixed]", "tx_tcp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "off [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "10.0.2.15", "broadcast": "10.0.2.255", "netmask": "255.255.255.0", "network": "10.0.2.0"}, "ipv6": [{"address": "fec0::5054:ff:fe12:3456", "prefix": "64", "scope": "site"}, {"address": "fe80::5054:ff:fe12:3456", "prefix": "64", "scope": "link"}], "macaddress": "52:54:00:12:34:56", "module": "virtio_net", "mtu": 1500, "pciid": "virtio0", "promisc": false, "speed": -1, "timestamping": ["tx_software", "rx_software", "software"], "type": "ether"}, "ansible_fibre_channel_wwn": [], "ansible_fips": false, "ansible_form_factor": "Other", "ansible_fqdn": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_hostname": "ibm-p8-kvm-03-guest-02", "ansible_hostnqn": "", "ansible_interfaces": ["eth0", "lo"], "ansible_is_chroot": false, "ansible_iscsi_iqn": "", "ansible_kernel": "4.18.0-80.el8.x86_64", "ansible_lo": {"active": true, "device": "lo", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "on", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "on [fixed]", "netns_local": "on [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "on [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "off [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "on", "tcp_segmentation_offload": "on", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "tx_checksumming": "on", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "on [fixed]", "tx_nocache_copy": "off [fixed]", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tx_sctp_segmentation": "on", "tx_tcp6_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp_segmentation": "on", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "on [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "127.0.0.1", "broadcast": "host", "netmask": "255.0.0.0", "network": "127.0.0.0"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "mtu": 65536, "promisc": false, "timestamping": ["tx_software", "rx_software", "software"], "type": "loopback"}, "ansible_local": {}, "ansible_lsb": {}, "ansible_machine": "x86_64", "ansible_machine_id": "72ddfb76328248ac8eaaff449aa907db", "ansible_memfree_mb": 387, "ansible_memory_mb": {"nocache": {"free": 625, "used": 195}, "real": {"free": 387, "total": 820, "used": 433}, "swap": {"cached": 0, "free": 0, "total": 0, "used": 0}}, "ansible_memtotal_mb": 820, "ansible_mounts": [{"block_available": 2290087, "block_size": 4096, "block_total": 2618619, "block_used": 328532, "device": "/dev/vda1", "fstype": "xfs", "inode_available": 5207720, "inode_total": 5242352, "inode_used": 34632, "mount": "/", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 9380196352, "size_total": 10725863424, "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}], "ansible_nodename": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_os_family": "RedHat", "ansible_pkg_mgr": "dnf", "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": ["ttyS0", "ttyS0,115200n8"], "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_processor": ["0", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "1", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "2", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz"], "ansible_processor_cores": 1, "ansible_processor_count": 3, "ansible_processor_threads_per_core": 1, "ansible_processor_vcpus": 3, "ansible_product_name": "Standard PC (i440FX + PIIX, 1996)", "ansible_product_serial": "NA", "ansible_product_uuid": "NA", "ansible_product_version": "pc-i440fx-3.1", "ansible_python": {"executable": "/usr/libexec/platform-python", "has_sslcontext": true, "type": "cpython", "version": {"major": 3, "micro": 8, "minor": 6, "releaselevel": "final", "serial": 0}, "version_info": [3, 6, 8, "final", 0]}, "ansible_python_version": "3.6.8", "ansible_real_group_id": 0, "ansible_real_user_id": 0, "ansible_selinux": {"config_mode": "enforcing", "mode": "enforcing", "policyvers": 31, "status": "enabled", "type": "targeted"}, "ansible_selinux_python_present": true, "ansible_service_mgr": "systemd", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBM1ZKlsASNMLWVAJyNZELEbjccFW1Y4Nfe7CK9E+FO9pKdsxSoow6V2wQhwNAU9paNkrqWCHY1KY7PCB+Xzg/1o=", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIALsboeHipL8AzNitI5gTtUPyLb2aPytf7B1+fndAN51", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABAQDJ8SXboNZxJSzsvHYgosvCpequ5K7RdsjjzloDenbswPatzCKJtmXTXNPZKCCmpbP13TJwgtpyZS/4u4U9eYJLSrgIbi1b+DBHEyaBL9SW55USq+vsPMoGJOvSNvbCoiDjhFl3QykN1TgOuTreqB5YhCFiBSFJDPiJ2KAAaAKmoDFAgfsZuQYwgvtMY3WrlVVfapQVMzv07H58YMhakrIBtOgPjp1Mn7Vm9ajEyIaV3GizBnI+8jb8hLyyriXckOzQt83xrbpChl3w+45J3g5Ph6t7+/IfDrcftrEMtoQcnVgjNNMwG6ECeXJcmmKZJMHe+ikKWKNB73oW6ev7+7eh", "ansible_swapfree_mb": 0, "ansible_swaptotal_mb": 0, "ansible_system": "Linux", "ansible_system_capabilities": ["cap_chown", "cap_dac_override", "cap_dac_read_search", "cap_fowner", "cap_fsetid", "cap_kill", "cap_setgid", "cap_setuid", "cap_setpcap", "cap_linux_immutable", "cap_net_bind_service", "cap_net_broadcast", "cap_net_admin", "cap_net_raw", "cap_ipc_lock", "cap_ipc_owner", "cap_sys_module", "cap_sys_rawio", "cap_sys_chroot", "cap_sys_ptrace", "cap_sys_pacct", "cap_sys_admin", "cap_sys_boot", "cap_sys_nice", "cap_sys_resource", "cap_sys_time", "cap_sys_tty_config", "cap_mknod", "cap_lease", "cap_audit_write", "cap_audit_control", "cap_setfcap", "cap_mac_override", "cap_mac_admin", "cap_syslog", "cap_wake_alarm", "cap_block_suspend", "cap_audit_read+ep"], "ansible_system_capabilities_enforced": "True", "ansible_system_vendor": "QEMU", "ansible_uptime_seconds": 97, "ansible_user_dir": "/root", "ansible_user_gecos": "root", "ansible_user_gid": 0, "ansible_user_id": "root", "ansible_user_shell": "/bin/bash", "ansible_user_uid": 0, "ansible_userspace_architecture": "x86_64", "ansible_userspace_bits": "64", "ansible_virtualization_role": "guest", "ansible_virtualization_type": "kvm", "gather_subset": ["all"], "module_setup": true}, "changed": false} META: ran handlers TASK [include_role : storage] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_change_mount.yml:10 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "msg": "", "rc": 0, "results": ["Installed: python3-blivet", "Installed: libblockdev-part-2.19-7.el8.x86_64", "Installed: nss-3.41.0-5.el8.x86_64", "Installed: libblockdev-swap-2.19-7.el8.x86_64", "Installed: userspace-rcu-0.10.1-2.el8.x86_64", "Installed: libblockdev-utils-2.19-7.el8.x86_64", "Installed: mdadm-4.1-4.el8.x86_64", "Installed: nss-softokn-3.41.0-5.el8.x86_64", "Installed: libbytesize-1.4-1.el8.x86_64", "Installed: nss-softokn-freebl-3.41.0-5.el8.x86_64", "Installed: nss-sysinit-3.41.0-5.el8.x86_64", "Installed: nss-util-3.41.0-5.el8.x86_64", "Installed: volume_key-libs-0.3.11-4.el8.x86_64", "Installed: ndctl-62-2.el8.x86_64", "Installed: ndctl-libs-62-2.el8.x86_64", "Installed: blivet-data-1:3.1.0-9.el8.noarch", "Installed: python3-pyparted-1:3.11.0-13.el8.x86_64", "Installed: python3-blivet-1:3.1.0-9.el8.noarch", "Installed: python3-blockdev-2.19-7.el8.x86_64", "Installed: daxctl-libs-62-2.el8.x86_64", "Installed: gdisk-1.0.3-6.el8.x86_64", "Installed: python3-bytesize-1.4-1.el8.x86_64", "Installed: device-mapper-event-8:1.02.155-6.el8.x86_64", "Installed: device-mapper-event-libs-8:1.02.155-6.el8.x86_64", "Installed: libaio-0.3.110-12.el8.x86_64", "Installed: device-mapper-multipath-0.7.8-7.el8.x86_64", "Installed: device-mapper-multipath-libs-0.7.8-7.el8.x86_64", "Installed: device-mapper-persistent-data-0.7.6-1.el8.x86_64", "Installed: lsof-4.91-2.el8.x86_64", "Installed: libblockdev-2.19-7.el8.x86_64", "Installed: lvm2-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-crypto-2.19-7.el8.x86_64", "Installed: libblockdev-dm-2.19-7.el8.x86_64", "Installed: lvm2-libs-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-fs-2.19-7.el8.x86_64", "Installed: libblockdev-kbd-2.19-7.el8.x86_64", "Installed: libblockdev-loop-2.19-7.el8.x86_64", "Installed: libblockdev-lvm-2.19-7.el8.x86_64", "Installed: libblockdev-mdraid-2.19-7.el8.x86_64", "Installed: libblockdev-mpath-2.19-7.el8.x86_64", "Installed: nspr-4.20.0-2.el8.x86_64", "Installed: libblockdev-nvdimm-2.19-7.el8.x86_64"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": [], "changed": false, "leaves": [], "mounts": [], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [], "changed": false, "failed": false, "leaves": [], "mounts": [], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_change_mount.yml:13 included: /tmp/tmpfvho35gc/tests/get_unused_disk.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Find unused disks in the system] ********************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "disks": ["vdb"]} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:8 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"unused_disks": ["vdb"]}, "changed": false} TASK [Exit playbook when there's no unused disks in the system] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:13 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} META: TASK [Print unused disks] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:19 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "unused_disks": [ "vdb" ] } TASK [Create a LVM logical volume mounted at "/opt/test1"] ************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/tests_change_mount.yml:18 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "msg": "Nothing to do", "rc": 0, "results": ["Installed: python3-blivet"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}]}]}, "ansible_loop_var": "pool", "changed": false, "pool": {"disks": ["vdb"], "name": "foo", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test1'}]}, {'name': 'test1', 'size': '3g', 'mount_point': '/opt/test1'}]) => {"ansible_facts": {"vol_defaults": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}], "vol_pools": ["foo"], "vols": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}]}, "ansible_loop_var": "item", "changed": false, "item": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}]}, {"mount_point": "/opt/test1", "name": "test1", "size": "3g"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test1'}, {'state': 'present', 'type': 'lvm', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid'}]) => {"ansible_facts": {"all_vols": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}]}, "ansible_loop_var": "item", "changed": false, "idx": 0, "item": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"vols_by_pool": {"foo": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}]}}, "ansible_loop_var": "item", "changed": false, "item": {"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}]}]}, "ansible_loop_var": "pool", "changed": false, "idx": 0, "pool": {"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}]}]}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [ { "disks": [ "vdb" ], "name": "foo", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" } ] } ] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": ["create format lvmpv on /dev/vdb", "create device /dev/foo", "create device /dev/mapper/foo-test1", "create format xfs on /dev/mapper/foo-test1"], "changed": true, "leaves": ["/dev/sr0", "/dev/vda1", "/dev/vdc", "/dev/mapper/foo-test1"], "mounts": [{"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted"}], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [ "create format lvmpv on /dev/vdb", "create device /dev/foo", "create device /dev/mapper/foo-test1", "create format xfs on /dev/mapper/foo-test1" ], "changed": true, "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vdc", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted"}, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1"} TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "cmd": ["systemctl", "daemon-reload"], "delta": "0:00:00.501633", "end": "2019-05-26 06:32:06.926423", "rc": 0, "start": "2019-05-26 06:32:06.424790", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_change_mount.yml:30 included: /tmp/tmpfvho35gc/tests/verify-role-results.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:1 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"storage_tests_list": ["verify-volumes"]}, "changed": false} TASK [Print out pool information] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:4 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "_storage_pools_list": [ { "disks": [ "vdb" ], "name": "foo", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" } ] } ] } TASK [Print out volume information] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:9 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} TASK [Ensure the specified devices in the role were correctly configured in the user's system] ************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:14 included: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Read the /etc/fstab file for volume existence] ******************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "cmd": ["cat", "/etc/fstab"], "delta": "0:00:00.012296", "end": "2019-05-26 06:32:11.060779", "rc": 0, "start": "2019-05-26 06:32:11.048483", "stderr": "", "stderr_lines": [], "stdout": "\n#\n# /etc/fstab\n# Created by anaconda on Thu Apr 4 08:39:20 2019\n#\n# Accessible filesystems, by reference, are maintained under '/dev/disk/'.\n# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.\n#\n# After editing this file, run 'systemctl daemon-reload' to update systemd\n# units generated from this file.\n#\nUUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0\n/dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0", "stdout_lines": ["", "#", "# /etc/fstab", "# Created by anaconda on Thu Apr 4 08:39:20 2019", "#", "# Accessible filesystems, by reference, are maintained under '/dev/disk/'.", "# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.", "#", "# After editing this file, run 'systemctl daemon-reload' to update systemd", "# units generated from this file.", "#", "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0", "/dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0"]} TASK [Verify the volumes listed in storage_pools were correctly managed] *********************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:7 included: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"mount_entries": []}, "changed": false} TASK [gather mounts related to each volume in the pool] **************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => {"ansible_facts": {"mount_entries": [[{"block_available": 770128, "block_size": 4096, "block_total": 783872, "block_used": 13744, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3154444288, "size_total": 3210739712, "uuid": "95508ffe-b155-4d3e-883b-7d8fc751d7ef"}]]}, "ansible_loop_var": "current_volume", "changed": false, "current_volume": {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}} TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:12 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:25 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab.] **************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:38 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, "msg": "All assertions passed" } TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:14 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test1", "device": "/dev/mapper/foo-test1"} TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:21 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vda1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vda1", "device": "/dev/vda1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test1", "device": "/dev/mapper/foo-test1"} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:28 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_mount_devs": ["/dev/vda1", "/dev/mapper/foo-test1"], "blivet_mount_devs": ["/dev/mapper/foo-test1"], "removed_volumes": []}, "changed": false} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:33 TASK [Verify the volumes with no pool were correctly managed] ********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:40 TASK [Verify the present volumes with no pool were correctly managed] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:52 TASK [Verify the absent volumes with no pool are not in ansible_mounts] ************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:65 TASK [Change the mount location to "/opt/test2"] *********************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_change_mount.yml:32 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "msg": "Nothing to do", "rc": 0, "results": ["Installed: python3-blivet"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test2'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test2", "name": "test1", "size": "3g"}]}]}, "ansible_loop_var": "pool", "changed": false, "pool": {"disks": ["vdb"], "name": "foo", "volumes": [{"mount_point": "/opt/test2", "name": "test1", "size": "3g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test2'}]}, {'name': 'test1', 'size': '3g', 'mount_point': '/opt/test2'}]) => {"ansible_facts": {"vol_defaults": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}], "vol_pools": ["foo"], "vols": [{"mount_point": "/opt/test2", "name": "test1", "size": "3g"}]}, "ansible_loop_var": "item", "changed": false, "item": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test2", "name": "test1", "size": "3g"}]}, {"mount_point": "/opt/test2", "name": "test1", "size": "3g"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test2'}, {'state': 'present', 'type': 'lvm', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid'}]) => {"ansible_facts": {"all_vols": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}]}, "ansible_loop_var": "item", "changed": false, "idx": 0, "item": [{"mount_point": "/opt/test2", "name": "test1", "size": "3g"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test2'}]}) => {"ansible_facts": {"vols_by_pool": {"foo": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}]}}, "ansible_loop_var": "item", "changed": false, "item": {"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test2", "name": "test1", "size": "3g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test2'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}]}]}, "ansible_loop_var": "pool", "changed": false, "idx": 0, "pool": {"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test2", "name": "test1", "size": "3g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}]}]}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [ { "disks": [ "vdb" ], "name": "foo", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" } ] } ] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": [], "changed": false, "leaves": ["/dev/sr0", "/dev/vda1", "/dev/mapper/foo-test1", "/dev/vdc"], "mounts": [{"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test1", "state": "mounted"}], "removed_mounts": [{"device": "/dev/mapper/foo-test1", "path": "/opt/test1"}]} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [], "changed": false, "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/mapper/foo-test1", "/dev/vdc" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "removed_mounts": [ { "device": "/dev/mapper/foo-test1", "path": "/opt/test1" } ] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'path': '/opt/test1', 'device': '/dev/mapper/foo-test1'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "mount_info": {"device": "/dev/mapper/foo-test1", "path": "/opt/test1"}, "name": "/opt/test1", "opts": "defaults", "passno": "0"} TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test2', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test1", "state": "mounted"}, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1"} TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "cmd": ["systemctl", "daemon-reload"], "delta": "0:00:00.509250", "end": "2019-05-26 06:32:29.655292", "rc": 0, "start": "2019-05-26 06:32:29.146042", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_change_mount.yml:44 included: /tmp/tmpfvho35gc/tests/verify-role-results.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:1 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"storage_tests_list": ["verify-volumes"]}, "changed": false} TASK [Print out pool information] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:4 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "_storage_pools_list": [ { "disks": [ "vdb" ], "name": "foo", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" } ] } ] } TASK [Print out volume information] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:9 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} TASK [Ensure the specified devices in the role were correctly configured in the user's system] ************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:14 included: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Read the /etc/fstab file for volume existence] ******************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "cmd": ["cat", "/etc/fstab"], "delta": "0:00:00.009603", "end": "2019-05-26 06:32:34.536601", "rc": 0, "start": "2019-05-26 06:32:34.526998", "stderr": "", "stderr_lines": [], "stdout": "\n#\n# /etc/fstab\n# Created by anaconda on Thu Apr 4 08:39:20 2019\n#\n# Accessible filesystems, by reference, are maintained under '/dev/disk/'.\n# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.\n#\n# After editing this file, run 'systemctl daemon-reload' to update systemd\n# units generated from this file.\n#\nUUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0\n/dev/mapper/foo-test1 /opt/test2 xfs defaults 0 0", "stdout_lines": ["", "#", "# /etc/fstab", "# Created by anaconda on Thu Apr 4 08:39:20 2019", "#", "# Accessible filesystems, by reference, are maintained under '/dev/disk/'.", "# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.", "#", "# After editing this file, run 'systemctl daemon-reload' to update systemd", "# units generated from this file.", "#", "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0", "/dev/mapper/foo-test1 /opt/test2 xfs defaults 0 0"]} TASK [Verify the volumes listed in storage_pools were correctly managed] *********************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:7 included: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"mount_entries": []}, "changed": false} TASK [gather mounts related to each volume in the pool] **************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => {"ansible_facts": {"mount_entries": [[{"block_available": 770128, "block_size": 4096, "block_total": 783872, "block_used": 13744, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3154444288, "size_total": 3210739712, "uuid": "95508ffe-b155-4d3e-883b-7d8fc751d7ef"}]]}, "ansible_loop_var": "current_volume", "changed": false, "current_volume": {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}} TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:12 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:25 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab.] **************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:38 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, "msg": "All assertions passed" } TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:14 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test1", "device": "/dev/mapper/foo-test1"} TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:21 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vda1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vda1", "device": "/dev/vda1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test1", "device": "/dev/mapper/foo-test1"} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:28 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_mount_devs": ["/dev/vda1", "/dev/mapper/foo-test1"], "blivet_mount_devs": ["/dev/mapper/foo-test1"], "removed_volumes": []}, "changed": false} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:33 TASK [Verify the volumes with no pool were correctly managed] ********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:40 TASK [Verify the present volumes with no pool were correctly managed] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:52 TASK [Verify the absent volumes with no pool are not in ansible_mounts] ************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:65 META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=86 changed=8 unreachable=0 failed=0 skipped=27 rescued=0 ignored=0 + cd /tmp/tmpfvho35gc/tests; TEST_SUBJECTS=/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TEST_ARTIFACTS=/tmp/linux-system-role-test-work-pull-linux-system-roles_storage-25-db52957-rhel-8-df0ousve/artifacts ansible-playbook -vv --inventory=/usr/share/ansible/inventory/standard-inventory-qcow2 /tmp/tmpfvho35gc/_setup.yml /tmp/tmpfvho35gc/tests/tests_create_disk_then_remove.yml ansible-playbook 2.8.0 config file = /home/till/.ansible.cfg configured module search path = ['/home/till/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 3.7.3 (default, May 11 2019, 00:38:04) [GCC 9.1.1 20190503 (Red Hat 9.1.1-1)] Using /home/till/.ansible.cfg as config file PLAYBOOK: _setup.yml *************************************************************************************************************************************************************************************** 2 plays in /tmp/tmpfvho35gc/_setup.yml PLAY [Fail when only localhost is available] *************************************************************************************************************************************************************** META: ran handlers TASK [debug] *********************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/_setup.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "groups": { "all": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "localhost": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "subjects": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "ungrouped": [] } } TASK [fail] ************************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:7 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} META: ran handlers META: ran handlers PLAY [Setup repos] ***************************************************************************************************************************************************************************************** META: ran handlers TASK [set up internal repositories] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:16 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true} META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=2 changed=1 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 PLAYBOOK: tests_create_disk_then_remove.yml **************************************************************************************************************************************************************** 1 plays in /tmp/tmpfvho35gc/tests/tests_create_disk_then_remove.yml PLAY [all] ************************************************************************************************************************************************************************************************* TASK [Gathering Facts] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/tests_create_disk_then_remove.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_all_ipv4_addresses": ["10.0.2.15"], "ansible_all_ipv6_addresses": ["fec0::5054:ff:fe12:3456", "fe80::5054:ff:fe12:3456"], "ansible_apparmor": {"status": "disabled"}, "ansible_architecture": "x86_64", "ansible_bios_date": "04/01/2014", "ansible_bios_version": "1.12.0-2.fc30", "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": "ttyS0,115200n8", "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_date_time": {"date": "2019-05-26", "day": "26", "epoch": "1558866864", "hour": "06", "iso8601": "2019-05-26T10:34:24Z", "iso8601_basic": "20190526T063424104056", "iso8601_basic_short": "20190526T063424", "iso8601_micro": "2019-05-26T10:34:24.104265Z", "minute": "34", "month": "05", "second": "24", "time": "06:34:24", "tz": "EDT", "tz_offset": "-0400", "weekday": "Sonntag", "weekday_number": "0", "weeknumber": "20", "year": "2019"}, "ansible_default_ipv4": {"address": "10.0.2.15", "alias": "eth0", "broadcast": "10.0.2.255", "gateway": "10.0.2.2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "netmask": "255.255.255.0", "network": "10.0.2.0", "type": "ether"}, "ansible_default_ipv6": {"address": "fec0::5054:ff:fe12:3456", "gateway": "fe80::2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "prefix": "64", "scope": "site", "type": "ether"}, "ansible_device_links": {"ids": {"sr0": ["ata-QEMU_DVD-ROM_QM00003"]}, "labels": {"sr0": ["cidata"]}, "masters": {}, "uuids": {"sr0": ["2019-05-26-12-32-40-00"], "vda1": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}}, "ansible_devices": {"sr0": {"holders": [], "host": "IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]", "links": {"ids": ["ata-QEMU_DVD-ROM_QM00003"], "labels": ["cidata"], "masters": [], "uuids": ["2019-05-26-12-32-40-00"]}, "model": "QEMU DVD-ROM", "partitions": {}, "removable": "1", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "728", "sectorsize": "2048", "size": "364.00 KB", "support_discard": "0", "vendor": "QEMU", "virtual": 1}, "vda": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {"vda1": {"holders": [], "links": {"ids": [], "labels": [], "masters": [], "uuids": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}, "sectors": "20969439", "sectorsize": 512, "size": "10.00 GB", "start": "2048", "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdb": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdc": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}}, "ansible_distribution": "RedHat", "ansible_distribution_file_parsed": true, "ansible_distribution_file_path": "/etc/redhat-release", "ansible_distribution_file_search_string": "Red Hat", "ansible_distribution_file_variety": "RedHat", "ansible_distribution_major_version": "8", "ansible_distribution_release": "Ootpa", "ansible_distribution_version": "8.0", "ansible_dns": {"nameservers": ["10.0.2.3"], "search": ["virt.pnr.lab.eng.rdu2.redhat.com"]}, "ansible_domain": "virt.pnr.lab.eng.rdu2.redhat.com", "ansible_effective_group_id": 0, "ansible_effective_user_id": 0, "ansible_env": {"DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "HOME": "/root", "LANG": "de_DE.UTF-8", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "LOGNAME": "root", "LS_COLORS": "rs=0:di=38;5;33:ln=38;5;51:mh=00:pi=40;38;5;11:so=38;5;13:do=38;5;5:bd=48;5;232;38;5;11:cd=48;5;232;38;5;3:or=48;5;232;38;5;9:mi=01;05;37;41:su=48;5;196;38;5;15:sg=48;5;11;38;5;16:ca=48;5;196;38;5;226:tw=48;5;10;38;5;16:ow=48;5;10;38;5;21:st=48;5;21;38;5;15:ex=38;5;40:*.tar=38;5;9:*.tgz=38;5;9:*.arc=38;5;9:*.arj=38;5;9:*.taz=38;5;9:*.lha=38;5;9:*.lz4=38;5;9:*.lzh=38;5;9:*.lzma=38;5;9:*.tlz=38;5;9:*.txz=38;5;9:*.tzo=38;5;9:*.t7z=38;5;9:*.zip=38;5;9:*.z=38;5;9:*.dz=38;5;9:*.gz=38;5;9:*.lrz=38;5;9:*.lz=38;5;9:*.lzo=38;5;9:*.xz=38;5;9:*.zst=38;5;9:*.tzst=38;5;9:*.bz2=38;5;9:*.bz=38;5;9:*.tbz=38;5;9:*.tbz2=38;5;9:*.tz=38;5;9:*.deb=38;5;9:*.rpm=38;5;9:*.jar=38;5;9:*.war=38;5;9:*.ear=38;5;9:*.sar=38;5;9:*.rar=38;5;9:*.alz=38;5;9:*.ace=38;5;9:*.zoo=38;5;9:*.cpio=38;5;9:*.7z=38;5;9:*.rz=38;5;9:*.cab=38;5;9:*.wim=38;5;9:*.swm=38;5;9:*.dwm=38;5;9:*.esd=38;5;9:*.jpg=38;5;13:*.jpeg=38;5;13:*.mjpg=38;5;13:*.mjpeg=38;5;13:*.gif=38;5;13:*.bmp=38;5;13:*.pbm=38;5;13:*.pgm=38;5;13:*.ppm=38;5;13:*.tga=38;5;13:*.xbm=38;5;13:*.xpm=38;5;13:*.tif=38;5;13:*.tiff=38;5;13:*.png=38;5;13:*.svg=38;5;13:*.svgz=38;5;13:*.mng=38;5;13:*.pcx=38;5;13:*.mov=38;5;13:*.mpg=38;5;13:*.mpeg=38;5;13:*.m2v=38;5;13:*.mkv=38;5;13:*.webm=38;5;13:*.ogm=38;5;13:*.mp4=38;5;13:*.m4v=38;5;13:*.mp4v=38;5;13:*.vob=38;5;13:*.qt=38;5;13:*.nuv=38;5;13:*.wmv=38;5;13:*.asf=38;5;13:*.rm=38;5;13:*.rmvb=38;5;13:*.flc=38;5;13:*.avi=38;5;13:*.fli=38;5;13:*.flv=38;5;13:*.gl=38;5;13:*.dl=38;5;13:*.xcf=38;5;13:*.xwd=38;5;13:*.yuv=38;5;13:*.cgm=38;5;13:*.emf=38;5;13:*.ogv=38;5;13:*.ogx=38;5;13:*.aac=38;5;45:*.au=38;5;45:*.flac=38;5;45:*.m4a=38;5;45:*.mid=38;5;45:*.midi=38;5;45:*.mka=38;5;45:*.mp3=38;5;45:*.mpc=38;5;45:*.ogg=38;5;45:*.ra=38;5;45:*.wav=38;5;45:*.oga=38;5;45:*.opus=38;5;45:*.spx=38;5;45:*.xspf=38;5;45:", "MAIL": "/var/mail/root", "PATH": "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "PWD": "/root", "SELINUX_LEVEL_REQUESTED": "", "SELINUX_ROLE_REQUESTED": "", "SELINUX_USE_CURRENT_RANGE": "", "SHELL": "/bin/bash", "SHLVL": "2", "SSH_CLIENT": "10.0.2.2 52590 22", "SSH_CONNECTION": "10.0.2.2 52590 10.0.2.15 22", "SSH_TTY": "/dev/pts/0", "TERM": "xterm-256color", "USER": "root", "XDG_RUNTIME_DIR": "/run/user/0", "XDG_SESSION_ID": "1", "_": "/usr/libexec/platform-python"}, "ansible_eth0": {"active": true, "device": "eth0", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "off [requested on]", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "off [fixed]", "netns_local": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "off [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "on [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "off", "tcp_segmentation_offload": "off", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "tx_checksumming": "off", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "off [fixed]", "tx_nocache_copy": "off", "tx_scatter_gather": "off [fixed]", "tx_scatter_gather_fraglist": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_tcp6_segmentation": "off [fixed]", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off [fixed]", "tx_tcp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "off [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "10.0.2.15", "broadcast": "10.0.2.255", "netmask": "255.255.255.0", "network": "10.0.2.0"}, "ipv6": [{"address": "fec0::5054:ff:fe12:3456", "prefix": "64", "scope": "site"}, {"address": "fe80::5054:ff:fe12:3456", "prefix": "64", "scope": "link"}], "macaddress": "52:54:00:12:34:56", "module": "virtio_net", "mtu": 1500, "pciid": "virtio0", "promisc": false, "speed": -1, "timestamping": ["tx_software", "rx_software", "software"], "type": "ether"}, "ansible_fibre_channel_wwn": [], "ansible_fips": false, "ansible_form_factor": "Other", "ansible_fqdn": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_hostname": "ibm-p8-kvm-03-guest-02", "ansible_hostnqn": "", "ansible_interfaces": ["lo", "eth0"], "ansible_is_chroot": false, "ansible_iscsi_iqn": "", "ansible_kernel": "4.18.0-80.el8.x86_64", "ansible_lo": {"active": true, "device": "lo", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "on", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "on [fixed]", "netns_local": "on [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "on [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "off [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "on", "tcp_segmentation_offload": "on", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "tx_checksumming": "on", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "on [fixed]", "tx_nocache_copy": "off [fixed]", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tx_sctp_segmentation": "on", "tx_tcp6_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp_segmentation": "on", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "on [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "127.0.0.1", "broadcast": "host", "netmask": "255.0.0.0", "network": "127.0.0.0"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "mtu": 65536, "promisc": false, "timestamping": ["tx_software", "rx_software", "software"], "type": "loopback"}, "ansible_local": {}, "ansible_lsb": {}, "ansible_machine": "x86_64", "ansible_machine_id": "19708761fd5d472c82e4c3f481d99df1", "ansible_memfree_mb": 385, "ansible_memory_mb": {"nocache": {"free": 623, "used": 197}, "real": {"free": 385, "total": 820, "used": 435}, "swap": {"cached": 0, "free": 0, "total": 0, "used": 0}}, "ansible_memtotal_mb": 820, "ansible_mounts": [{"block_available": 2290089, "block_size": 4096, "block_total": 2618619, "block_used": 328530, "device": "/dev/vda1", "fstype": "xfs", "inode_available": 5207720, "inode_total": 5242352, "inode_used": 34632, "mount": "/", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 9380204544, "size_total": 10725863424, "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}], "ansible_nodename": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_os_family": "RedHat", "ansible_pkg_mgr": "dnf", "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": ["ttyS0", "ttyS0,115200n8"], "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_processor": ["0", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "1", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "2", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz"], "ansible_processor_cores": 1, "ansible_processor_count": 3, "ansible_processor_threads_per_core": 1, "ansible_processor_vcpus": 3, "ansible_product_name": "Standard PC (i440FX + PIIX, 1996)", "ansible_product_serial": "NA", "ansible_product_uuid": "NA", "ansible_product_version": "pc-i440fx-3.1", "ansible_python": {"executable": "/usr/libexec/platform-python", "has_sslcontext": true, "type": "cpython", "version": {"major": 3, "micro": 8, "minor": 6, "releaselevel": "final", "serial": 0}, "version_info": [3, 6, 8, "final", 0]}, "ansible_python_version": "3.6.8", "ansible_real_group_id": 0, "ansible_real_user_id": 0, "ansible_selinux": {"config_mode": "enforcing", "mode": "enforcing", "policyvers": 31, "status": "enabled", "type": "targeted"}, "ansible_selinux_python_present": true, "ansible_service_mgr": "systemd", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBJNGkJcM6aKRK1wh45lmDPEMOTuH96NPCbQ+TrEjhWe6D6XFk90fz9ML3ktWgPt1Qs3B0VLWtRAo9ehwAiiuDcA=", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIJM8WiP0OGp1Ic0jMiupGvCYExEsAMxgJVJWvjxX6TgS", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABAQDpsHof9s/Fxk7vei8fB31NwCmPTroaLJd7dXk2pYXC3mg583jN9754BDECdR18PdsiyfeMS4Pi7HJEirg3q5qa371CNQaKq7Im1w5LPAiHDqEsmAISEfdRHzfDdc2SS/vGDvc7eMEVQewSwzIstHg57XXIVEFJ1UXhlk80Bn08qJLAbrCyHFw5kQ4paj/5Luq+FbvEIWWIMbWSkaArNO0ApXd/P+0G2aEPOgjuDddEP2KCEZdZVN9EQA3umw8sxkT3/Np1y+uLJ03wjDl0JtnYPKDrUuMcWYrGpx4YvCMEdBGD+dGMvpAjKU1LfuJ0+TOVzmruILR9fBigI/LFF53N", "ansible_swapfree_mb": 0, "ansible_swaptotal_mb": 0, "ansible_system": "Linux", "ansible_system_capabilities": ["cap_chown", "cap_dac_override", "cap_dac_read_search", "cap_fowner", "cap_fsetid", "cap_kill", "cap_setgid", "cap_setuid", "cap_setpcap", "cap_linux_immutable", "cap_net_bind_service", "cap_net_broadcast", "cap_net_admin", "cap_net_raw", "cap_ipc_lock", "cap_ipc_owner", "cap_sys_module", "cap_sys_rawio", "cap_sys_chroot", "cap_sys_ptrace", "cap_sys_pacct", "cap_sys_admin", "cap_sys_boot", "cap_sys_nice", "cap_sys_resource", "cap_sys_time", "cap_sys_tty_config", "cap_mknod", "cap_lease", "cap_audit_write", "cap_audit_control", "cap_setfcap", "cap_mac_override", "cap_mac_admin", "cap_syslog", "cap_wake_alarm", "cap_block_suspend", "cap_audit_read+ep"], "ansible_system_capabilities_enforced": "True", "ansible_system_vendor": "QEMU", "ansible_uptime_seconds": 98, "ansible_user_dir": "/root", "ansible_user_gecos": "root", "ansible_user_gid": 0, "ansible_user_id": "root", "ansible_user_shell": "/bin/bash", "ansible_user_uid": 0, "ansible_userspace_architecture": "x86_64", "ansible_userspace_bits": "64", "ansible_virtualization_role": "guest", "ansible_virtualization_type": "kvm", "gather_subset": ["all"], "module_setup": true}, "changed": false} META: ran handlers TASK [include_role : storage] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_create_disk_then_remove.yml:8 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "msg": "", "rc": 0, "results": ["Installed: python3-blivet", "Installed: libblockdev-part-2.19-7.el8.x86_64", "Installed: nss-3.41.0-5.el8.x86_64", "Installed: libblockdev-swap-2.19-7.el8.x86_64", "Installed: userspace-rcu-0.10.1-2.el8.x86_64", "Installed: libblockdev-utils-2.19-7.el8.x86_64", "Installed: mdadm-4.1-4.el8.x86_64", "Installed: nss-softokn-3.41.0-5.el8.x86_64", "Installed: libbytesize-1.4-1.el8.x86_64", "Installed: nss-softokn-freebl-3.41.0-5.el8.x86_64", "Installed: nss-sysinit-3.41.0-5.el8.x86_64", "Installed: nss-util-3.41.0-5.el8.x86_64", "Installed: volume_key-libs-0.3.11-4.el8.x86_64", "Installed: ndctl-62-2.el8.x86_64", "Installed: ndctl-libs-62-2.el8.x86_64", "Installed: blivet-data-1:3.1.0-9.el8.noarch", "Installed: python3-pyparted-1:3.11.0-13.el8.x86_64", "Installed: python3-blivet-1:3.1.0-9.el8.noarch", "Installed: python3-blockdev-2.19-7.el8.x86_64", "Installed: daxctl-libs-62-2.el8.x86_64", "Installed: gdisk-1.0.3-6.el8.x86_64", "Installed: python3-bytesize-1.4-1.el8.x86_64", "Installed: device-mapper-event-8:1.02.155-6.el8.x86_64", "Installed: device-mapper-event-libs-8:1.02.155-6.el8.x86_64", "Installed: libaio-0.3.110-12.el8.x86_64", "Installed: device-mapper-multipath-0.7.8-7.el8.x86_64", "Installed: device-mapper-multipath-libs-0.7.8-7.el8.x86_64", "Installed: device-mapper-persistent-data-0.7.6-1.el8.x86_64", "Installed: lsof-4.91-2.el8.x86_64", "Installed: libblockdev-2.19-7.el8.x86_64", "Installed: lvm2-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-crypto-2.19-7.el8.x86_64", "Installed: libblockdev-dm-2.19-7.el8.x86_64", "Installed: lvm2-libs-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-fs-2.19-7.el8.x86_64", "Installed: libblockdev-kbd-2.19-7.el8.x86_64", "Installed: libblockdev-loop-2.19-7.el8.x86_64", "Installed: libblockdev-lvm-2.19-7.el8.x86_64", "Installed: libblockdev-mdraid-2.19-7.el8.x86_64", "Installed: libblockdev-mpath-2.19-7.el8.x86_64", "Installed: nspr-4.20.0-2.el8.x86_64", "Installed: libblockdev-nvdimm-2.19-7.el8.x86_64"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": [], "changed": false, "leaves": [], "mounts": [], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [], "changed": false, "failed": false, "leaves": [], "mounts": [], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_create_disk_then_remove.yml:11 included: /tmp/tmpfvho35gc/tests/get_unused_disk.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Find unused disks in the system] ********************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:2 [WARNING]: The value 0 (type int) in a string field was converted to '0' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change. ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "disks": ["vdb"]} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:8 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"unused_disks": ["vdb"]}, "changed": false} TASK [Exit playbook when there's no unused disks in the system] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:13 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} META: TASK [Print unused disks] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:19 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "unused_disks": [ "vdb" ] } TASK [Create a disk device mounted on "/opt/test1"] ******************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_create_disk_then_remove.yml:15 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "msg": "Nothing to do", "rc": 0, "results": ["Installed: python3-blivet"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'name': 'test1', 'type': 'disk', 'disks': ['vdb'], 'mount_point': '/opt/test1'}) => {"ansible_facts": {"volumes": [{"disks": ["vdb"], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "present", "type": "disk"}]}, "ansible_loop_var": "volume", "changed": false, "volume": {"disks": ["vdb"], "mount_point": "/opt/test1", "name": "test1", "type": "disk"}} TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": [{"disks": ["vdb"], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "present", "type": "disk"}]}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [ { "disks": [ "vdb" ], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "present", "type": "disk" } ] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": ["create format xfs on /dev/vdb"], "changed": true, "leaves": ["/dev/sr0", "/dev/vda1", "/dev/vdb", "/dev/vdc"], "mounts": [{"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/vdb", "state": "mounted"}], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [ "create format xfs on /dev/vdb" ], "changed": true, "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vdb", "/dev/vdc" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/vdb", "state": "mounted" } ], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'src': '/dev/vdb', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/vdb", "state": "mounted"}, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/vdb"} TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "cmd": ["systemctl", "daemon-reload"], "delta": "0:00:00.599056", "end": "2019-05-26 06:35:35.815426", "rc": 0, "start": "2019-05-26 06:35:35.216370", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_create_disk_then_remove.yml:25 included: /tmp/tmpfvho35gc/tests/verify-role-results.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:1 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"storage_tests_list": ["verify-volumes"]}, "changed": false} TASK [Print out pool information] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:4 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} TASK [Print out volume information] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:9 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "_storage_volumes_list": [ { "disks": [ "vdb" ], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "present", "type": "disk" } ] } TASK [Ensure the specified devices in the role were correctly configured in the user's system] ************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:14 included: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Read the /etc/fstab file for volume existence] ******************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "cmd": ["cat", "/etc/fstab"], "delta": "0:00:00.003147", "end": "2019-05-26 06:35:40.455238", "rc": 0, "start": "2019-05-26 06:35:40.452091", "stderr": "", "stderr_lines": [], "stdout": "\n#\n# /etc/fstab\n# Created by anaconda on Thu Apr 4 08:39:20 2019\n#\n# Accessible filesystems, by reference, are maintained under '/dev/disk/'.\n# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.\n#\n# After editing this file, run 'systemctl daemon-reload' to update systemd\n# units generated from this file.\n#\nUUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0\n/dev/vdb /opt/test1 xfs defaults 0 0", "stdout_lines": ["", "#", "# /etc/fstab", "# Created by anaconda on Thu Apr 4 08:39:20 2019", "#", "# Accessible filesystems, by reference, are maintained under '/dev/disk/'.", "# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.", "#", "# After editing this file, run 'systemctl daemon-reload' to update systemd", "# units generated from this file.", "#", "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0", "/dev/vdb /opt/test1 xfs defaults 0 0"]} TASK [Verify the volumes listed in storage_pools were correctly managed] *********************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:7 TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:14 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vdb) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vdb", "device": "/dev/vdb"} TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:21 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vda1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vda1", "device": "/dev/vda1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vdb) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vdb", "device": "/dev/vdb"} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:28 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_mount_devs": ["/dev/vda1", "/dev/vdb"], "blivet_mount_devs": ["/dev/vdb"], "removed_volumes": []}, "changed": false} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:33 TASK [Verify the volumes with no pool were correctly managed] ********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:40 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'disk', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'disks': ['vdb']}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "disks": [ "vdb" ], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "present", "type": "disk" }, "msg": "All assertions passed" } TASK [Verify the present volumes with no pool were correctly managed] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:52 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'disk', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'disks': ['vdb']}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "disks": [ "vdb" ], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "present", "type": "disk" }, "idx": 0, "msg": "All assertions passed" } TASK [Verify the absent volumes with no pool are not in ansible_mounts] ************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:65 TASK [Remove the disk device created above] **************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_create_disk_then_remove.yml:27 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "msg": "Nothing to do", "rc": 0, "results": ["Installed: python3-blivet"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'name': 'test1', 'type': 'disk', 'disks': ['vdb'], 'mount_point': '/opt/test1', 'state': 'absent'}) => {"ansible_facts": {"volumes": [{"disks": ["vdb"], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "absent", "type": "disk"}]}, "ansible_loop_var": "volume", "changed": false, "volume": {"disks": ["vdb"], "mount_point": "/opt/test1", "name": "test1", "state": "absent", "type": "disk"}} TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": [{"disks": ["vdb"], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "absent", "type": "disk"}]}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [ { "disks": [ "vdb" ], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "absent", "type": "disk" } ] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": ["destroy format xfs on /dev/vdb"], "changed": true, "leaves": ["/dev/sr0", "/dev/vda1", "/dev/vdb", "/dev/vdc"], "mounts": [], "removed_mounts": [{"device": "/dev/vdb", "path": "/opt/test1"}]} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [ "destroy format xfs on /dev/vdb" ], "changed": true, "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vdb", "/dev/vdc" ], "mounts": [], "removed_mounts": [ { "device": "/dev/vdb", "path": "/opt/test1" } ] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'path': '/opt/test1', 'device': '/dev/vdb'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "mount_info": {"device": "/dev/vdb", "path": "/opt/test1"}, "name": "/opt/test1", "opts": "defaults", "passno": "0"} TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "cmd": ["systemctl", "daemon-reload"], "delta": "0:00:00.589028", "end": "2019-05-26 06:35:59.855137", "rc": 0, "start": "2019-05-26 06:35:59.266109", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_create_disk_then_remove.yml:38 included: /tmp/tmpfvho35gc/tests/verify-role-results.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:1 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"storage_tests_list": ["verify-volumes"]}, "changed": false} TASK [Print out pool information] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:4 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} TASK [Print out volume information] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:9 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "_storage_volumes_list": [ { "disks": [ "vdb" ], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "absent", "type": "disk" } ] } TASK [Ensure the specified devices in the role were correctly configured in the user's system] ************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:14 included: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Read the /etc/fstab file for volume existence] ******************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "cmd": ["cat", "/etc/fstab"], "delta": "0:00:00.013520", "end": "2019-05-26 06:36:04.247715", "rc": 0, "start": "2019-05-26 06:36:04.234195", "stderr": "", "stderr_lines": [], "stdout": "\n#\n# /etc/fstab\n# Created by anaconda on Thu Apr 4 08:39:20 2019\n#\n# Accessible filesystems, by reference, are maintained under '/dev/disk/'.\n# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.\n#\n# After editing this file, run 'systemctl daemon-reload' to update systemd\n# units generated from this file.\n#\nUUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0", "stdout_lines": ["", "#", "# /etc/fstab", "# Created by anaconda on Thu Apr 4 08:39:20 2019", "#", "# Accessible filesystems, by reference, are maintained under '/dev/disk/'.", "# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.", "#", "# After editing this file, run 'systemctl daemon-reload' to update systemd", "# units generated from this file.", "#", "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0"]} TASK [Verify the volumes listed in storage_pools were correctly managed] *********************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:7 TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:14 TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:21 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vda1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vda1", "device": "/dev/vda1"} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:28 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_mount_devs": ["/dev/vda1"], "blivet_mount_devs": [], "removed_volumes": []}, "changed": false} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:33 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'absent', 'type': 'disk', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'disks': ['vdb']}) => {"ansible_facts": {"removed_volumes": [{"_device": {"device": "/dev/vdb", "path": "/opt/test1"}, "disks": ["vdb"], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "absent", "type": "disk"}]}, "ansible_loop_var": "current_volume", "changed": false, "current_volume": {"disks": ["vdb"], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "absent", "type": "disk"}} TASK [Verify the volumes with no pool were correctly managed] ********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:40 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'absent', 'type': 'disk', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'disks': ['vdb']}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "disks": [ "vdb" ], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "absent", "type": "disk" }, "msg": "All assertions passed" } TASK [Verify the present volumes with no pool were correctly managed] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:52 TASK [Verify the absent volumes with no pool are not in ansible_mounts] ************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:65 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'absent', 'type': 'disk', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'disks': ['vdb'], '_device': {'path': '/opt/test1', 'device': '/dev/vdb'}}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "_device": { "device": "/dev/vdb", "path": "/opt/test1" }, "disks": [ "vdb" ], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "absent", "type": "disk" }, "msg": "All assertions passed" } META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=69 changed=8 unreachable=0 failed=0 skipped=34 rescued=0 ignored=0 + cd /tmp/tmpfvho35gc/tests; TEST_SUBJECTS=/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TEST_ARTIFACTS=/tmp/linux-system-role-test-work-pull-linux-system-roles_storage-25-db52957-rhel-8-df0ousve/artifacts ansible-playbook -vv --inventory=/usr/share/ansible/inventory/standard-inventory-qcow2 /tmp/tmpfvho35gc/_setup.yml /tmp/tmpfvho35gc/tests/tests_create_lvm_pool_then_remove.yml ansible-playbook 2.8.0 config file = /home/till/.ansible.cfg configured module search path = ['/home/till/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 3.7.3 (default, May 11 2019, 00:38:04) [GCC 9.1.1 20190503 (Red Hat 9.1.1-1)] Using /home/till/.ansible.cfg as config file PLAYBOOK: _setup.yml *************************************************************************************************************************************************************************************** 2 plays in /tmp/tmpfvho35gc/_setup.yml PLAY [Fail when only localhost is available] *************************************************************************************************************************************************************** META: ran handlers TASK [debug] *********************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/_setup.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "groups": { "all": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "localhost": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "subjects": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "ungrouped": [] } } TASK [fail] ************************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:7 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} META: ran handlers META: ran handlers PLAY [Setup repos] ***************************************************************************************************************************************************************************************** META: ran handlers TASK [set up internal repositories] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:16 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true} META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=2 changed=1 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 PLAYBOOK: tests_create_lvm_pool_then_remove.yml ************************************************************************************************************************************************************ 1 plays in /tmp/tmpfvho35gc/tests/tests_create_lvm_pool_then_remove.yml PLAY [all] ************************************************************************************************************************************************************************************************* TASK [Gathering Facts] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/tests_create_lvm_pool_then_remove.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_all_ipv4_addresses": ["10.0.2.15"], "ansible_all_ipv6_addresses": ["fec0::5054:ff:fe12:3456", "fe80::5054:ff:fe12:3456"], "ansible_apparmor": {"status": "disabled"}, "ansible_architecture": "x86_64", "ansible_bios_date": "04/01/2014", "ansible_bios_version": "1.12.0-2.fc30", "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": "ttyS0,115200n8", "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_date_time": {"date": "2019-05-26", "day": "26", "epoch": "1558867069", "hour": "06", "iso8601": "2019-05-26T10:37:49Z", "iso8601_basic": "20190526T063749071919", "iso8601_basic_short": "20190526T063749", "iso8601_micro": "2019-05-26T10:37:49.072045Z", "minute": "37", "month": "05", "second": "49", "time": "06:37:49", "tz": "EDT", "tz_offset": "-0400", "weekday": "Sonntag", "weekday_number": "0", "weeknumber": "20", "year": "2019"}, "ansible_default_ipv4": {"address": "10.0.2.15", "alias": "eth0", "broadcast": "10.0.2.255", "gateway": "10.0.2.2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "netmask": "255.255.255.0", "network": "10.0.2.0", "type": "ether"}, "ansible_default_ipv6": {"address": "fec0::5054:ff:fe12:3456", "gateway": "fe80::2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "prefix": "64", "scope": "site", "type": "ether"}, "ansible_device_links": {"ids": {"sr0": ["ata-QEMU_DVD-ROM_QM00003"]}, "labels": {"sr0": ["cidata"]}, "masters": {}, "uuids": {"sr0": ["2019-05-26-12-36-07-00"], "vda1": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}}, "ansible_devices": {"sr0": {"holders": [], "host": "IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]", "links": {"ids": ["ata-QEMU_DVD-ROM_QM00003"], "labels": ["cidata"], "masters": [], "uuids": ["2019-05-26-12-36-07-00"]}, "model": "QEMU DVD-ROM", "partitions": {}, "removable": "1", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "728", "sectorsize": "2048", "size": "364.00 KB", "support_discard": "0", "vendor": "QEMU", "virtual": 1}, "vda": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {"vda1": {"holders": [], "links": {"ids": [], "labels": [], "masters": [], "uuids": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}, "sectors": "20969439", "sectorsize": 512, "size": "10.00 GB", "start": "2048", "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdb": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdc": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}}, "ansible_distribution": "RedHat", "ansible_distribution_file_parsed": true, "ansible_distribution_file_path": "/etc/redhat-release", "ansible_distribution_file_search_string": "Red Hat", "ansible_distribution_file_variety": "RedHat", "ansible_distribution_major_version": "8", "ansible_distribution_release": "Ootpa", "ansible_distribution_version": "8.0", "ansible_dns": {"nameservers": ["10.0.2.3"], "search": ["virt.pnr.lab.eng.rdu2.redhat.com"]}, "ansible_domain": "virt.pnr.lab.eng.rdu2.redhat.com", "ansible_effective_group_id": 0, "ansible_effective_user_id": 0, "ansible_env": {"DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "HOME": "/root", "LANG": "de_DE.UTF-8", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "LOGNAME": "root", "LS_COLORS": "rs=0:di=38;5;33:ln=38;5;51:mh=00:pi=40;38;5;11:so=38;5;13:do=38;5;5:bd=48;5;232;38;5;11:cd=48;5;232;38;5;3:or=48;5;232;38;5;9:mi=01;05;37;41:su=48;5;196;38;5;15:sg=48;5;11;38;5;16:ca=48;5;196;38;5;226:tw=48;5;10;38;5;16:ow=48;5;10;38;5;21:st=48;5;21;38;5;15:ex=38;5;40:*.tar=38;5;9:*.tgz=38;5;9:*.arc=38;5;9:*.arj=38;5;9:*.taz=38;5;9:*.lha=38;5;9:*.lz4=38;5;9:*.lzh=38;5;9:*.lzma=38;5;9:*.tlz=38;5;9:*.txz=38;5;9:*.tzo=38;5;9:*.t7z=38;5;9:*.zip=38;5;9:*.z=38;5;9:*.dz=38;5;9:*.gz=38;5;9:*.lrz=38;5;9:*.lz=38;5;9:*.lzo=38;5;9:*.xz=38;5;9:*.zst=38;5;9:*.tzst=38;5;9:*.bz2=38;5;9:*.bz=38;5;9:*.tbz=38;5;9:*.tbz2=38;5;9:*.tz=38;5;9:*.deb=38;5;9:*.rpm=38;5;9:*.jar=38;5;9:*.war=38;5;9:*.ear=38;5;9:*.sar=38;5;9:*.rar=38;5;9:*.alz=38;5;9:*.ace=38;5;9:*.zoo=38;5;9:*.cpio=38;5;9:*.7z=38;5;9:*.rz=38;5;9:*.cab=38;5;9:*.wim=38;5;9:*.swm=38;5;9:*.dwm=38;5;9:*.esd=38;5;9:*.jpg=38;5;13:*.jpeg=38;5;13:*.mjpg=38;5;13:*.mjpeg=38;5;13:*.gif=38;5;13:*.bmp=38;5;13:*.pbm=38;5;13:*.pgm=38;5;13:*.ppm=38;5;13:*.tga=38;5;13:*.xbm=38;5;13:*.xpm=38;5;13:*.tif=38;5;13:*.tiff=38;5;13:*.png=38;5;13:*.svg=38;5;13:*.svgz=38;5;13:*.mng=38;5;13:*.pcx=38;5;13:*.mov=38;5;13:*.mpg=38;5;13:*.mpeg=38;5;13:*.m2v=38;5;13:*.mkv=38;5;13:*.webm=38;5;13:*.ogm=38;5;13:*.mp4=38;5;13:*.m4v=38;5;13:*.mp4v=38;5;13:*.vob=38;5;13:*.qt=38;5;13:*.nuv=38;5;13:*.wmv=38;5;13:*.asf=38;5;13:*.rm=38;5;13:*.rmvb=38;5;13:*.flc=38;5;13:*.avi=38;5;13:*.fli=38;5;13:*.flv=38;5;13:*.gl=38;5;13:*.dl=38;5;13:*.xcf=38;5;13:*.xwd=38;5;13:*.yuv=38;5;13:*.cgm=38;5;13:*.emf=38;5;13:*.ogv=38;5;13:*.ogx=38;5;13:*.aac=38;5;45:*.au=38;5;45:*.flac=38;5;45:*.m4a=38;5;45:*.mid=38;5;45:*.midi=38;5;45:*.mka=38;5;45:*.mp3=38;5;45:*.mpc=38;5;45:*.ogg=38;5;45:*.ra=38;5;45:*.wav=38;5;45:*.oga=38;5;45:*.opus=38;5;45:*.spx=38;5;45:*.xspf=38;5;45:", "MAIL": "/var/mail/root", "PATH": "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "PWD": "/root", "SELINUX_LEVEL_REQUESTED": "", "SELINUX_ROLE_REQUESTED": "", "SELINUX_USE_CURRENT_RANGE": "", "SHELL": "/bin/bash", "SHLVL": "2", "SSH_CLIENT": "10.0.2.2 42530 22", "SSH_CONNECTION": "10.0.2.2 42530 10.0.2.15 22", "SSH_TTY": "/dev/pts/0", "TERM": "xterm-256color", "USER": "root", "XDG_RUNTIME_DIR": "/run/user/0", "XDG_SESSION_ID": "1", "_": "/usr/libexec/platform-python"}, "ansible_eth0": {"active": true, "device": "eth0", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "off [requested on]", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "off [fixed]", "netns_local": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "off [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "on [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "off", "tcp_segmentation_offload": "off", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "tx_checksumming": "off", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "off [fixed]", "tx_nocache_copy": "off", "tx_scatter_gather": "off [fixed]", "tx_scatter_gather_fraglist": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_tcp6_segmentation": "off [fixed]", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off [fixed]", "tx_tcp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "off [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "10.0.2.15", "broadcast": "10.0.2.255", "netmask": "255.255.255.0", "network": "10.0.2.0"}, "ipv6": [{"address": "fec0::5054:ff:fe12:3456", "prefix": "64", "scope": "site"}, {"address": "fe80::5054:ff:fe12:3456", "prefix": "64", "scope": "link"}], "macaddress": "52:54:00:12:34:56", "module": "virtio_net", "mtu": 1500, "pciid": "virtio0", "promisc": false, "speed": -1, "timestamping": ["tx_software", "rx_software", "software"], "type": "ether"}, "ansible_fibre_channel_wwn": [], "ansible_fips": false, "ansible_form_factor": "Other", "ansible_fqdn": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_hostname": "ibm-p8-kvm-03-guest-02", "ansible_hostnqn": "", "ansible_interfaces": ["eth0", "lo"], "ansible_is_chroot": false, "ansible_iscsi_iqn": "", "ansible_kernel": "4.18.0-80.el8.x86_64", "ansible_lo": {"active": true, "device": "lo", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "on", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "on [fixed]", "netns_local": "on [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "on [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "off [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "on", "tcp_segmentation_offload": "on", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "tx_checksumming": "on", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "on [fixed]", "tx_nocache_copy": "off [fixed]", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tx_sctp_segmentation": "on", "tx_tcp6_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp_segmentation": "on", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "on [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "127.0.0.1", "broadcast": "host", "netmask": "255.0.0.0", "network": "127.0.0.0"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "mtu": 65536, "promisc": false, "timestamping": ["tx_software", "rx_software", "software"], "type": "loopback"}, "ansible_local": {}, "ansible_lsb": {}, "ansible_machine": "x86_64", "ansible_machine_id": "5d1f0b0d39bd4e5888141a7a4b705c98", "ansible_memfree_mb": 388, "ansible_memory_mb": {"nocache": {"free": 626, "used": 194}, "real": {"free": 388, "total": 820, "used": 432}, "swap": {"cached": 0, "free": 0, "total": 0, "used": 0}}, "ansible_memtotal_mb": 820, "ansible_mounts": [{"block_available": 2290090, "block_size": 4096, "block_total": 2618619, "block_used": 328529, "device": "/dev/vda1", "fstype": "xfs", "inode_available": 5207720, "inode_total": 5242352, "inode_used": 34632, "mount": "/", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 9380208640, "size_total": 10725863424, "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}], "ansible_nodename": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_os_family": "RedHat", "ansible_pkg_mgr": "dnf", "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": ["ttyS0", "ttyS0,115200n8"], "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_processor": ["0", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "1", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "2", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz"], "ansible_processor_cores": 1, "ansible_processor_count": 3, "ansible_processor_threads_per_core": 1, "ansible_processor_vcpus": 3, "ansible_product_name": "Standard PC (i440FX + PIIX, 1996)", "ansible_product_serial": "NA", "ansible_product_uuid": "NA", "ansible_product_version": "pc-i440fx-3.1", "ansible_python": {"executable": "/usr/libexec/platform-python", "has_sslcontext": true, "type": "cpython", "version": {"major": 3, "micro": 8, "minor": 6, "releaselevel": "final", "serial": 0}, "version_info": [3, 6, 8, "final", 0]}, "ansible_python_version": "3.6.8", "ansible_real_group_id": 0, "ansible_real_user_id": 0, "ansible_selinux": {"config_mode": "enforcing", "mode": "enforcing", "policyvers": 31, "status": "enabled", "type": "targeted"}, "ansible_selinux_python_present": true, "ansible_service_mgr": "systemd", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBHXGkK8d7788Y1lpIzI/ucad3UPhNwOmMTKFAZTNgknBLSJU3Vvl4TNGqcoC/MmXcR9KuZeUak+pAF87NgTbz3A=", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIEXL5qElIO3/AFO4dKD/LU5p7mrs/S7tTH3AOT1dwlTE", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABAQDBgXP3hlVuVJUapTfbnTfGy+xVZ3CsuqRSM4QkyHwCu3yT2BiWAI+ivlPmrM31XOdTc85JqQeSocN3Rhse9TqqU1OEaFEmqQGnqtH2qzlPEW3XIqE4QNIcLK1w3loGiPgMsZtJZz4iLq3ofyS7L/FUAeuNoRr3G5hvNilus+hgeifRLnU2+iiyNKB69DkWwScM0SM9Yu1arPaYF1vh4+425GFdHFVlpMq2M98D30vNehoyk/Pud7ElPaefE6hopxO8EnGnxB3Jxcir0YL6C1Nv9mh2PApN54ZWI5gs1qDUGUX6Q4GKkMzgMjx+oYjCV+YqaT6EhdrgyC7jeSv1jGRr", "ansible_swapfree_mb": 0, "ansible_swaptotal_mb": 0, "ansible_system": "Linux", "ansible_system_capabilities": ["cap_chown", "cap_dac_override", "cap_dac_read_search", "cap_fowner", "cap_fsetid", "cap_kill", "cap_setgid", "cap_setuid", "cap_setpcap", "cap_linux_immutable", "cap_net_bind_service", "cap_net_broadcast", "cap_net_admin", "cap_net_raw", "cap_ipc_lock", "cap_ipc_owner", "cap_sys_module", "cap_sys_rawio", "cap_sys_chroot", "cap_sys_ptrace", "cap_sys_pacct", "cap_sys_admin", "cap_sys_boot", "cap_sys_nice", "cap_sys_resource", "cap_sys_time", "cap_sys_tty_config", "cap_mknod", "cap_lease", "cap_audit_write", "cap_audit_control", "cap_setfcap", "cap_mac_override", "cap_mac_admin", "cap_syslog", "cap_wake_alarm", "cap_block_suspend", "cap_audit_read+ep"], "ansible_system_capabilities_enforced": "True", "ansible_system_vendor": "QEMU", "ansible_uptime_seconds": 96, "ansible_user_dir": "/root", "ansible_user_gecos": "root", "ansible_user_gid": 0, "ansible_user_id": "root", "ansible_user_shell": "/bin/bash", "ansible_user_uid": 0, "ansible_userspace_architecture": "x86_64", "ansible_userspace_bits": "64", "ansible_virtualization_role": "guest", "ansible_virtualization_type": "kvm", "gather_subset": ["all"], "module_setup": true}, "changed": false} META: ran handlers TASK [include_role : storage] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_create_lvm_pool_then_remove.yml:12 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "msg": "", "rc": 0, "results": ["Installed: python3-blivet", "Installed: libblockdev-part-2.19-7.el8.x86_64", "Installed: nss-3.41.0-5.el8.x86_64", "Installed: libblockdev-swap-2.19-7.el8.x86_64", "Installed: userspace-rcu-0.10.1-2.el8.x86_64", "Installed: libblockdev-utils-2.19-7.el8.x86_64", "Installed: mdadm-4.1-4.el8.x86_64", "Installed: nss-softokn-3.41.0-5.el8.x86_64", "Installed: libbytesize-1.4-1.el8.x86_64", "Installed: nss-softokn-freebl-3.41.0-5.el8.x86_64", "Installed: nss-sysinit-3.41.0-5.el8.x86_64", "Installed: nss-util-3.41.0-5.el8.x86_64", "Installed: volume_key-libs-0.3.11-4.el8.x86_64", "Installed: ndctl-62-2.el8.x86_64", "Installed: ndctl-libs-62-2.el8.x86_64", "Installed: blivet-data-1:3.1.0-9.el8.noarch", "Installed: python3-pyparted-1:3.11.0-13.el8.x86_64", "Installed: python3-blivet-1:3.1.0-9.el8.noarch", "Installed: python3-blockdev-2.19-7.el8.x86_64", "Installed: daxctl-libs-62-2.el8.x86_64", "Installed: gdisk-1.0.3-6.el8.x86_64", "Installed: python3-bytesize-1.4-1.el8.x86_64", "Installed: device-mapper-event-8:1.02.155-6.el8.x86_64", "Installed: device-mapper-event-libs-8:1.02.155-6.el8.x86_64", "Installed: libaio-0.3.110-12.el8.x86_64", "Installed: device-mapper-multipath-0.7.8-7.el8.x86_64", "Installed: device-mapper-multipath-libs-0.7.8-7.el8.x86_64", "Installed: device-mapper-persistent-data-0.7.6-1.el8.x86_64", "Installed: lsof-4.91-2.el8.x86_64", "Installed: libblockdev-2.19-7.el8.x86_64", "Installed: lvm2-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-crypto-2.19-7.el8.x86_64", "Installed: libblockdev-dm-2.19-7.el8.x86_64", "Installed: lvm2-libs-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-fs-2.19-7.el8.x86_64", "Installed: libblockdev-kbd-2.19-7.el8.x86_64", "Installed: libblockdev-loop-2.19-7.el8.x86_64", "Installed: libblockdev-lvm-2.19-7.el8.x86_64", "Installed: libblockdev-mdraid-2.19-7.el8.x86_64", "Installed: libblockdev-mpath-2.19-7.el8.x86_64", "Installed: nspr-4.20.0-2.el8.x86_64", "Installed: libblockdev-nvdimm-2.19-7.el8.x86_64"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": [], "changed": false, "leaves": [], "mounts": [], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [], "changed": false, "failed": false, "leaves": [], "mounts": [], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_create_lvm_pool_then_remove.yml:15 included: /tmp/tmpfvho35gc/tests/get_unused_disk.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Find unused disks in the system] ********************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "disks": ["vdb"]} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:8 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"unused_disks": ["vdb"]}, "changed": false} TASK [Exit playbook when there's no unused disks in the system] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:13 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} META: TASK [Print unused disks] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:19 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "unused_disks": [ "vdb" ] } TASK [Create two LVM logical volumes under volume group 'foo'] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/tests_create_lvm_pool_then_remove.yml:20 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "msg": "Nothing to do", "rc": 0, "results": ["Installed: python3-blivet"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}]}, "ansible_loop_var": "pool", "changed": false, "pool": {"disks": ["vdb"], "name": "foo", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]}, {'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}]) => {"ansible_facts": {"vol_defaults": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}], "vol_pools": ["foo"], "vols": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}, "ansible_loop_var": "item", "changed": false, "item": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}, {"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]) => {"ansible_facts": {"vol_defaults": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}], "vol_pools": ["foo", "foo"], "vols": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}, "ansible_loop_var": "item", "changed": false, "item": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}, {'state': 'present', 'type': 'lvm', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid'}]) => {"ansible_facts": {"all_vols": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}]}, "ansible_loop_var": "item", "changed": false, "idx": 0, "item": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}]} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}, {'state': 'present', 'type': 'lvm', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid'}]) => {"ansible_facts": {"all_vols": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm"}]}, "ansible_loop_var": "item", "changed": false, "idx": 1, "item": [{"mount_point": "/opt/test2", "name": "test2", "size": "4g"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]}) => {"ansible_facts": {"vols_by_pool": {"foo": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm"}]}}, "ansible_loop_var": "item", "changed": false, "item": {"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm"}]}]}, "ansible_loop_var": "pool", "changed": false, "idx": 0, "pool": {"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm"}]}]}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [ { "disks": [ "vdb" ], "name": "foo", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" }, { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm" } ] } ] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": ["create format lvmpv on /dev/vdb", "create device /dev/foo", "create device /dev/mapper/foo-test2", "create format xfs on /dev/mapper/foo-test2", "create device /dev/mapper/foo-test1", "create format xfs on /dev/mapper/foo-test1"], "changed": true, "leaves": ["/dev/sr0", "/dev/vda1", "/dev/vdc", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2"], "mounts": [{"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted"}, {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted"}], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [ "create format lvmpv on /dev/vdb", "create device /dev/foo", "create device /dev/mapper/foo-test2", "create format xfs on /dev/mapper/foo-test2", "create device /dev/mapper/foo-test1", "create format xfs on /dev/mapper/foo-test1" ], "changed": true, "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vdc", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" } ], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted"}, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1"} changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'src': '/dev/mapper/foo-test2', 'path': '/opt/test2', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted"}, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2"} TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "cmd": ["systemctl", "daemon-reload"], "delta": "0:00:00.682373", "end": "2019-05-26 06:38:45.174403", "rc": 0, "start": "2019-05-26 06:38:44.492030", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_create_lvm_pool_then_remove.yml:35 included: /tmp/tmpfvho35gc/tests/verify-role-results.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:1 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"storage_tests_list": ["verify-volumes"]}, "changed": false} TASK [Print out pool information] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:4 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "_storage_pools_list": [ { "disks": [ "vdb" ], "name": "foo", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" }, { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm" } ] } ] } TASK [Print out volume information] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:9 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} TASK [Ensure the specified devices in the role were correctly configured in the user's system] ************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:14 included: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Read the /etc/fstab file for volume existence] ******************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "cmd": ["cat", "/etc/fstab"], "delta": "0:00:00.015947", "end": "2019-05-26 06:38:49.178325", "rc": 0, "start": "2019-05-26 06:38:49.162378", "stderr": "", "stderr_lines": [], "stdout": "\n#\n# /etc/fstab\n# Created by anaconda on Thu Apr 4 08:39:20 2019\n#\n# Accessible filesystems, by reference, are maintained under '/dev/disk/'.\n# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.\n#\n# After editing this file, run 'systemctl daemon-reload' to update systemd\n# units generated from this file.\n#\nUUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0\n/dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0\n/dev/mapper/foo-test2 /opt/test2 xfs defaults 0 0", "stdout_lines": ["", "#", "# /etc/fstab", "# Created by anaconda on Thu Apr 4 08:39:20 2019", "#", "# Accessible filesystems, by reference, are maintained under '/dev/disk/'.", "# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.", "#", "# After editing this file, run 'systemctl daemon-reload' to update systemd", "# units generated from this file.", "#", "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0", "/dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0", "/dev/mapper/foo-test2 /opt/test2 xfs defaults 0 0"]} TASK [Verify the volumes listed in storage_pools were correctly managed] *********************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:7 included: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"mount_entries": []}, "changed": false} TASK [gather mounts related to each volume in the pool] **************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => {"ansible_facts": {"mount_entries": [[{"block_available": 1290756, "block_size": 4096, "block_total": 1308160, "block_used": 17404, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2621437, "inode_total": 2621440, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 5286936576, "size_total": 5358223360, "uuid": "95f8898c-4d14-4b9d-a354-6d065ad9b34b"}]]}, "ansible_loop_var": "current_volume", "changed": false, "current_volume": {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test2', 'pool': 'foo'}) => {"ansible_facts": {"mount_entries": [[{"block_available": 1290756, "block_size": 4096, "block_total": 1308160, "block_used": 17404, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2621437, "inode_total": 2621440, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 5286936576, "size_total": 5358223360, "uuid": "95f8898c-4d14-4b9d-a354-6d065ad9b34b"}], [{"block_available": 1030436, "block_size": 4096, "block_total": 1046016, "block_used": 15580, "device": "/dev/mapper/foo-test2", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4220665856, "size_total": 4284481536, "uuid": "f318f26e-3a73-411c-8d5f-38db5e970d3f"}]]}, "ansible_loop_var": "current_volume", "changed": false, "current_volume": {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm"}} TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:12 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test2', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm" }, "idx": 1, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:25 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test2', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm" }, "idx": 1, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab.] **************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:38 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" }, "msg": "All assertions passed" } ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test2', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm" }, "msg": "All assertions passed" } TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:14 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test1", "device": "/dev/mapper/foo-test1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test2) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test2", "device": "/dev/mapper/foo-test2"} TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:21 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vda1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vda1", "device": "/dev/vda1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test1", "device": "/dev/mapper/foo-test1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test2) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test2", "device": "/dev/mapper/foo-test2"} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:28 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_mount_devs": ["/dev/vda1", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2"], "blivet_mount_devs": ["/dev/mapper/foo-test1", "/dev/mapper/foo-test2"], "removed_volumes": []}, "changed": false} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:33 TASK [Verify the volumes with no pool were correctly managed] ********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:40 TASK [Verify the present volumes with no pool were correctly managed] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:52 TASK [Verify the absent volumes with no pool are not in ansible_mounts] ************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:65 TASK [Remove both of the LVM logical volumes in 'foo' created above] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_create_lvm_pool_then_remove.yml:37 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "msg": "Nothing to do", "rc": 0, "results": ["Installed: python3-blivet"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'name': 'foo', 'disks': ['vdb'], 'state': 'absent', 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "foo", "state": "absent", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}]}, "ansible_loop_var": "pool", "changed": false, "pool": {"disks": ["vdb"], "name": "foo", "state": "absent", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'state': 'absent', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]}, {'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}]) => {"ansible_facts": {"vol_defaults": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}], "vol_pools": ["foo"], "vols": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]}, "ansible_loop_var": "item", "changed": false, "item": [{"disks": ["vdb"], "name": "foo", "state": "absent", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}, {"mount_point": "/opt/test1", "name": "test1", "size": "5g"}]} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'state': 'absent', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]) => {"ansible_facts": {"vol_defaults": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}], "vol_pools": ["foo", "foo"], "vols": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}, "ansible_loop_var": "item", "changed": false, "item": [{"disks": ["vdb"], "name": "foo", "state": "absent", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}, {'state': 'present', 'type': 'lvm', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid'}]) => {"ansible_facts": {"all_vols": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}]}, "ansible_loop_var": "item", "changed": false, "idx": 0, "item": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}]} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}, {'state': 'present', 'type': 'lvm', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid'}]) => {"ansible_facts": {"all_vols": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm"}]}, "ansible_loop_var": "item", "changed": false, "idx": 1, "item": [{"mount_point": "/opt/test2", "name": "test2", "size": "4g"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'absent', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]}) => {"ansible_facts": {"vols_by_pool": {"foo": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm"}]}}, "ansible_loop_var": "item", "changed": false, "item": {"disks": ["vdb"], "name": "foo", "state": "absent", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'absent', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '5g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "foo", "state": "absent", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm"}]}]}, "ansible_loop_var": "pool", "changed": false, "idx": 0, "pool": {"disks": ["vdb"], "name": "foo", "state": "absent", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "5g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": [{"disks": ["vdb"], "name": "foo", "state": "absent", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm"}]}]}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [ { "disks": [ "vdb" ], "name": "foo", "state": "absent", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" }, { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm" } ] } ] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": ["destroy format xfs on /dev/mapper/foo-test1", "destroy device /dev/mapper/foo-test1", "destroy format xfs on /dev/mapper/foo-test2", "destroy device /dev/mapper/foo-test2", "destroy device /dev/foo"], "changed": true, "leaves": ["/dev/sr0", "/dev/vda1", "/dev/vdb", "/dev/vdc"], "mounts": [], "removed_mounts": [{"device": "/dev/mapper/foo-test1", "path": "/opt/test1"}, {"device": "/dev/mapper/foo-test2", "path": "/opt/test2"}]} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [ "destroy format xfs on /dev/mapper/foo-test1", "destroy device /dev/mapper/foo-test1", "destroy format xfs on /dev/mapper/foo-test2", "destroy device /dev/mapper/foo-test2", "destroy device /dev/foo" ], "changed": true, "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vdb", "/dev/vdc" ], "mounts": [], "removed_mounts": [ { "device": "/dev/mapper/foo-test1", "path": "/opt/test1" }, { "device": "/dev/mapper/foo-test2", "path": "/opt/test2" } ] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'path': '/opt/test1', 'device': '/dev/mapper/foo-test1'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "mount_info": {"device": "/dev/mapper/foo-test1", "path": "/opt/test1"}, "name": "/opt/test1", "opts": "defaults", "passno": "0"} changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'path': '/opt/test2', 'device': '/dev/mapper/foo-test2'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "mount_info": {"device": "/dev/mapper/foo-test2", "path": "/opt/test2"}, "name": "/opt/test2", "opts": "defaults", "passno": "0"} TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "cmd": ["systemctl", "daemon-reload"], "delta": "0:00:00.632911", "end": "2019-05-26 06:39:17.682570", "rc": 0, "start": "2019-05-26 06:39:17.049659", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_create_lvm_pool_then_remove.yml:53 included: /tmp/tmpfvho35gc/tests/verify-role-results.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:1 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"storage_tests_list": ["verify-volumes"]}, "changed": false} TASK [Print out pool information] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:4 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "_storage_pools_list": [ { "disks": [ "vdb" ], "name": "foo", "state": "absent", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" }, { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm" } ] } ] } TASK [Print out volume information] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:9 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} TASK [Ensure the specified devices in the role were correctly configured in the user's system] ************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:14 included: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Read the /etc/fstab file for volume existence] ******************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "cmd": ["cat", "/etc/fstab"], "delta": "0:00:00.005983", "end": "2019-05-26 06:39:27.084709", "rc": 0, "start": "2019-05-26 06:39:27.078726", "stderr": "", "stderr_lines": [], "stdout": "\n#\n# /etc/fstab\n# Created by anaconda on Thu Apr 4 08:39:20 2019\n#\n# Accessible filesystems, by reference, are maintained under '/dev/disk/'.\n# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.\n#\n# After editing this file, run 'systemctl daemon-reload' to update systemd\n# units generated from this file.\n#\nUUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0", "stdout_lines": ["", "#", "# /etc/fstab", "# Created by anaconda on Thu Apr 4 08:39:20 2019", "#", "# Accessible filesystems, by reference, are maintained under '/dev/disk/'.", "# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.", "#", "# After editing this file, run 'systemctl daemon-reload' to update systemd", "# units generated from this file.", "#", "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0"]} TASK [Verify the volumes listed in storage_pools were correctly managed] *********************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:7 included: /tmp/tmpfvho35gc/tests/test-volumes-absent-helper.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Make sure all volumes in pool.volumes are absent] **************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-absent-helper.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '5g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "5g", "state": "present", "type": "lvm" }, "msg": "All assertions passed" } ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test2', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "4g", "state": "present", "type": "lvm" }, "msg": "All assertions passed" } TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:14 TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:21 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vda1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vda1", "device": "/dev/vda1"} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:28 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_mount_devs": ["/dev/vda1"], "blivet_mount_devs": [], "removed_volumes": []}, "changed": false} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:33 TASK [Verify the volumes with no pool were correctly managed] ********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:40 TASK [Verify the present volumes with no pool were correctly managed] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:52 TASK [Verify the absent volumes with no pool are not in ansible_mounts] ************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:65 META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=80 changed=8 unreachable=0 failed=0 skipped=29 rescued=0 ignored=0 + cd /tmp/tmpfvho35gc/tests; TEST_SUBJECTS=/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TEST_ARTIFACTS=/tmp/linux-system-role-test-work-pull-linux-system-roles_storage-25-db52957-rhel-8-df0ousve/artifacts ansible-playbook -vv --inventory=/usr/share/ansible/inventory/standard-inventory-qcow2 /tmp/tmpfvho35gc/_setup.yml /tmp/tmpfvho35gc/tests/tests_filesystem_one_disk.yml ansible-playbook 2.8.0 config file = /home/till/.ansible.cfg configured module search path = ['/home/till/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 3.7.3 (default, May 11 2019, 00:38:04) [GCC 9.1.1 20190503 (Red Hat 9.1.1-1)] Using /home/till/.ansible.cfg as config file PLAYBOOK: _setup.yml *************************************************************************************************************************************************************************************** 2 plays in /tmp/tmpfvho35gc/_setup.yml PLAY [Fail when only localhost is available] *************************************************************************************************************************************************************** META: ran handlers TASK [debug] *********************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/_setup.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "groups": { "all": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "localhost": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "subjects": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "ungrouped": [] } } TASK [fail] ************************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:7 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} META: ran handlers META: ran handlers PLAY [Setup repos] ***************************************************************************************************************************************************************************************** META: ran handlers TASK [set up internal repositories] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:16 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true} META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=2 changed=1 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 PLAYBOOK: tests_filesystem_one_disk.yml ******************************************************************************************************************************************************************** 1 plays in /tmp/tmpfvho35gc/tests/tests_filesystem_one_disk.yml PLAY [all] ************************************************************************************************************************************************************************************************* TASK [Gathering Facts] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/tests_filesystem_one_disk.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_all_ipv4_addresses": ["10.0.2.15"], "ansible_all_ipv6_addresses": ["fec0::5054:ff:fe12:3456", "fe80::5054:ff:fe12:3456"], "ansible_apparmor": {"status": "disabled"}, "ansible_architecture": "x86_64", "ansible_bios_date": "04/01/2014", "ansible_bios_version": "1.12.0-2.fc30", "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": "ttyS0,115200n8", "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_date_time": {"date": "2019-05-26", "day": "26", "epoch": "1558867269", "hour": "06", "iso8601": "2019-05-26T10:41:09Z", "iso8601_basic": "20190526T064109336377", "iso8601_basic_short": "20190526T064109", "iso8601_micro": "2019-05-26T10:41:09.336518Z", "minute": "41", "month": "05", "second": "09", "time": "06:41:09", "tz": "EDT", "tz_offset": "-0400", "weekday": "Sonntag", "weekday_number": "0", "weeknumber": "20", "year": "2019"}, "ansible_default_ipv4": {"address": "10.0.2.15", "alias": "eth0", "broadcast": "10.0.2.255", "gateway": "10.0.2.2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "netmask": "255.255.255.0", "network": "10.0.2.0", "type": "ether"}, "ansible_default_ipv6": {"address": "fec0::5054:ff:fe12:3456", "gateway": "fe80::2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "prefix": "64", "scope": "site", "type": "ether"}, "ansible_device_links": {"ids": {"sr0": ["ata-QEMU_DVD-ROM_QM00003"]}, "labels": {"sr0": ["cidata"]}, "masters": {}, "uuids": {"sr0": ["2019-05-26-12-39-30-00"], "vda1": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}}, "ansible_devices": {"sr0": {"holders": [], "host": "IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]", "links": {"ids": ["ata-QEMU_DVD-ROM_QM00003"], "labels": ["cidata"], "masters": [], "uuids": ["2019-05-26-12-39-30-00"]}, "model": "QEMU DVD-ROM", "partitions": {}, "removable": "1", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "728", "sectorsize": "2048", "size": "364.00 KB", "support_discard": "0", "vendor": "QEMU", "virtual": 1}, "vda": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {"vda1": {"holders": [], "links": {"ids": [], "labels": [], "masters": [], "uuids": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}, "sectors": "20969439", "sectorsize": 512, "size": "10.00 GB", "start": "2048", "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdb": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdc": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}}, "ansible_distribution": "RedHat", "ansible_distribution_file_parsed": true, "ansible_distribution_file_path": "/etc/redhat-release", "ansible_distribution_file_search_string": "Red Hat", "ansible_distribution_file_variety": "RedHat", "ansible_distribution_major_version": "8", "ansible_distribution_release": "Ootpa", "ansible_distribution_version": "8.0", "ansible_dns": {"nameservers": ["10.0.2.3"], "search": ["virt.pnr.lab.eng.rdu2.redhat.com"]}, "ansible_domain": "virt.pnr.lab.eng.rdu2.redhat.com", "ansible_effective_group_id": 0, "ansible_effective_user_id": 0, "ansible_env": {"DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "HOME": "/root", "LANG": "de_DE.UTF-8", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "LOGNAME": "root", "LS_COLORS": "rs=0:di=38;5;33:ln=38;5;51:mh=00:pi=40;38;5;11:so=38;5;13:do=38;5;5:bd=48;5;232;38;5;11:cd=48;5;232;38;5;3:or=48;5;232;38;5;9:mi=01;05;37;41:su=48;5;196;38;5;15:sg=48;5;11;38;5;16:ca=48;5;196;38;5;226:tw=48;5;10;38;5;16:ow=48;5;10;38;5;21:st=48;5;21;38;5;15:ex=38;5;40:*.tar=38;5;9:*.tgz=38;5;9:*.arc=38;5;9:*.arj=38;5;9:*.taz=38;5;9:*.lha=38;5;9:*.lz4=38;5;9:*.lzh=38;5;9:*.lzma=38;5;9:*.tlz=38;5;9:*.txz=38;5;9:*.tzo=38;5;9:*.t7z=38;5;9:*.zip=38;5;9:*.z=38;5;9:*.dz=38;5;9:*.gz=38;5;9:*.lrz=38;5;9:*.lz=38;5;9:*.lzo=38;5;9:*.xz=38;5;9:*.zst=38;5;9:*.tzst=38;5;9:*.bz2=38;5;9:*.bz=38;5;9:*.tbz=38;5;9:*.tbz2=38;5;9:*.tz=38;5;9:*.deb=38;5;9:*.rpm=38;5;9:*.jar=38;5;9:*.war=38;5;9:*.ear=38;5;9:*.sar=38;5;9:*.rar=38;5;9:*.alz=38;5;9:*.ace=38;5;9:*.zoo=38;5;9:*.cpio=38;5;9:*.7z=38;5;9:*.rz=38;5;9:*.cab=38;5;9:*.wim=38;5;9:*.swm=38;5;9:*.dwm=38;5;9:*.esd=38;5;9:*.jpg=38;5;13:*.jpeg=38;5;13:*.mjpg=38;5;13:*.mjpeg=38;5;13:*.gif=38;5;13:*.bmp=38;5;13:*.pbm=38;5;13:*.pgm=38;5;13:*.ppm=38;5;13:*.tga=38;5;13:*.xbm=38;5;13:*.xpm=38;5;13:*.tif=38;5;13:*.tiff=38;5;13:*.png=38;5;13:*.svg=38;5;13:*.svgz=38;5;13:*.mng=38;5;13:*.pcx=38;5;13:*.mov=38;5;13:*.mpg=38;5;13:*.mpeg=38;5;13:*.m2v=38;5;13:*.mkv=38;5;13:*.webm=38;5;13:*.ogm=38;5;13:*.mp4=38;5;13:*.m4v=38;5;13:*.mp4v=38;5;13:*.vob=38;5;13:*.qt=38;5;13:*.nuv=38;5;13:*.wmv=38;5;13:*.asf=38;5;13:*.rm=38;5;13:*.rmvb=38;5;13:*.flc=38;5;13:*.avi=38;5;13:*.fli=38;5;13:*.flv=38;5;13:*.gl=38;5;13:*.dl=38;5;13:*.xcf=38;5;13:*.xwd=38;5;13:*.yuv=38;5;13:*.cgm=38;5;13:*.emf=38;5;13:*.ogv=38;5;13:*.ogx=38;5;13:*.aac=38;5;45:*.au=38;5;45:*.flac=38;5;45:*.m4a=38;5;45:*.mid=38;5;45:*.midi=38;5;45:*.mka=38;5;45:*.mp3=38;5;45:*.mpc=38;5;45:*.ogg=38;5;45:*.ra=38;5;45:*.wav=38;5;45:*.oga=38;5;45:*.opus=38;5;45:*.spx=38;5;45:*.xspf=38;5;45:", "MAIL": "/var/mail/root", "PATH": "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "PWD": "/root", "SELINUX_LEVEL_REQUESTED": "", "SELINUX_ROLE_REQUESTED": "", "SELINUX_USE_CURRENT_RANGE": "", "SHELL": "/bin/bash", "SHLVL": "2", "SSH_CLIENT": "10.0.2.2 38960 22", "SSH_CONNECTION": "10.0.2.2 38960 10.0.2.15 22", "SSH_TTY": "/dev/pts/0", "TERM": "xterm-256color", "USER": "root", "XDG_RUNTIME_DIR": "/run/user/0", "XDG_SESSION_ID": "1", "_": "/usr/libexec/platform-python"}, "ansible_eth0": {"active": true, "device": "eth0", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "off [requested on]", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "off [fixed]", "netns_local": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "off [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "on [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "off", "tcp_segmentation_offload": "off", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "tx_checksumming": "off", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "off [fixed]", "tx_nocache_copy": "off", "tx_scatter_gather": "off [fixed]", "tx_scatter_gather_fraglist": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_tcp6_segmentation": "off [fixed]", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off [fixed]", "tx_tcp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "off [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "10.0.2.15", "broadcast": "10.0.2.255", "netmask": "255.255.255.0", "network": "10.0.2.0"}, "ipv6": [{"address": "fec0::5054:ff:fe12:3456", "prefix": "64", "scope": "site"}, {"address": "fe80::5054:ff:fe12:3456", "prefix": "64", "scope": "link"}], "macaddress": "52:54:00:12:34:56", "module": "virtio_net", "mtu": 1500, "pciid": "virtio0", "promisc": false, "speed": -1, "timestamping": ["tx_software", "rx_software", "software"], "type": "ether"}, "ansible_fibre_channel_wwn": [], "ansible_fips": false, "ansible_form_factor": "Other", "ansible_fqdn": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_hostname": "ibm-p8-kvm-03-guest-02", "ansible_hostnqn": "", "ansible_interfaces": ["eth0", "lo"], "ansible_is_chroot": false, "ansible_iscsi_iqn": "", "ansible_kernel": "4.18.0-80.el8.x86_64", "ansible_lo": {"active": true, "device": "lo", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "on", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "on [fixed]", "netns_local": "on [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "on [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "off [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "on", "tcp_segmentation_offload": "on", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "tx_checksumming": "on", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "on [fixed]", "tx_nocache_copy": "off [fixed]", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tx_sctp_segmentation": "on", "tx_tcp6_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp_segmentation": "on", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "on [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "127.0.0.1", "broadcast": "host", "netmask": "255.0.0.0", "network": "127.0.0.0"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "mtu": 65536, "promisc": false, "timestamping": ["tx_software", "rx_software", "software"], "type": "loopback"}, "ansible_local": {}, "ansible_lsb": {}, "ansible_machine": "x86_64", "ansible_machine_id": "9ca2c13b8ba64514969a6d2253c38612", "ansible_memfree_mb": 394, "ansible_memory_mb": {"nocache": {"free": 626, "used": 194}, "real": {"free": 394, "total": 820, "used": 426}, "swap": {"cached": 0, "free": 0, "total": 0, "used": 0}}, "ansible_memtotal_mb": 820, "ansible_mounts": [{"block_available": 2291652, "block_size": 4096, "block_total": 2618619, "block_used": 326967, "device": "/dev/vda1", "fstype": "xfs", "inode_available": 5207721, "inode_total": 5242352, "inode_used": 34631, "mount": "/", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 9386606592, "size_total": 10725863424, "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}], "ansible_nodename": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_os_family": "RedHat", "ansible_pkg_mgr": "dnf", "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": ["ttyS0", "ttyS0,115200n8"], "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_processor": ["0", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "1", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "2", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz"], "ansible_processor_cores": 1, "ansible_processor_count": 3, "ansible_processor_threads_per_core": 1, "ansible_processor_vcpus": 3, "ansible_product_name": "Standard PC (i440FX + PIIX, 1996)", "ansible_product_serial": "NA", "ansible_product_uuid": "NA", "ansible_product_version": "pc-i440fx-3.1", "ansible_python": {"executable": "/usr/libexec/platform-python", "has_sslcontext": true, "type": "cpython", "version": {"major": 3, "micro": 8, "minor": 6, "releaselevel": "final", "serial": 0}, "version_info": [3, 6, 8, "final", 0]}, "ansible_python_version": "3.6.8", "ansible_real_group_id": 0, "ansible_real_user_id": 0, "ansible_selinux": {"config_mode": "enforcing", "mode": "enforcing", "policyvers": 31, "status": "enabled", "type": "targeted"}, "ansible_selinux_python_present": true, "ansible_service_mgr": "systemd", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBN4V+yXTjn7dHyZQIIRM2QvMhCJs9wuXIvpbcehUfcM3RvmGilJLITEozvIUhxNYWxeEBqypGxVQ99UAj4Uj8ss=", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIBsXISjrTeUMpa9KKcuyYR3gVCQA6E6S12i8HHTjbpSi", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABAQDJ2G/f+Z3n6qL0GDAOq2099qmOjoUt3gEnPXs9yzUoX2QnEQ1IWVMIiofX1f1/jfvHHJY+RBPOS+0Ehinn/e5+0RgONdI6+eNYeZ2h5DJ3TZtrMgGN0mEeY1LTc6Tgf1IH4xRnytp4425xCxQo5he81z8gNC+3CYySTB6Ab0k0saQVFC2GS0IhyvnQfBeKaodf1zc+276WP708XXVY8NN1v8D+P9SG8lu5K4a+f7gkVQQavEiPBkg5bOOfbNegLkY6IM9V2Yi6h3nNvBhmrNy1zj+VLh9gTAPfCFENaBEfOhUAdIXQW7YK4u9iFTuy7przoCDREmDspd3gvkBuiNrX", "ansible_swapfree_mb": 0, "ansible_swaptotal_mb": 0, "ansible_system": "Linux", "ansible_system_capabilities": ["cap_chown", "cap_dac_override", "cap_dac_read_search", "cap_fowner", "cap_fsetid", "cap_kill", "cap_setgid", "cap_setuid", "cap_setpcap", "cap_linux_immutable", "cap_net_bind_service", "cap_net_broadcast", "cap_net_admin", "cap_net_raw", "cap_ipc_lock", "cap_ipc_owner", "cap_sys_module", "cap_sys_rawio", "cap_sys_chroot", "cap_sys_ptrace", "cap_sys_pacct", "cap_sys_admin", "cap_sys_boot", "cap_sys_nice", "cap_sys_resource", "cap_sys_time", "cap_sys_tty_config", "cap_mknod", "cap_lease", "cap_audit_write", "cap_audit_control", "cap_setfcap", "cap_mac_override", "cap_mac_admin", "cap_syslog", "cap_wake_alarm", "cap_block_suspend", "cap_audit_read+ep"], "ansible_system_capabilities_enforced": "True", "ansible_system_vendor": "QEMU", "ansible_uptime_seconds": 92, "ansible_user_dir": "/root", "ansible_user_gecos": "root", "ansible_user_gid": 0, "ansible_user_id": "root", "ansible_user_shell": "/bin/bash", "ansible_user_uid": 0, "ansible_userspace_architecture": "x86_64", "ansible_userspace_bits": "64", "ansible_virtualization_role": "guest", "ansible_virtualization_type": "kvm", "gather_subset": ["all"], "module_setup": true}, "changed": false} META: ran handlers TASK [include_role : storage] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_filesystem_one_disk.yml:8 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "msg": "", "rc": 0, "results": ["Installed: python3-blivet", "Installed: libblockdev-part-2.19-7.el8.x86_64", "Installed: nss-3.41.0-5.el8.x86_64", "Installed: libblockdev-swap-2.19-7.el8.x86_64", "Installed: userspace-rcu-0.10.1-2.el8.x86_64", "Installed: libblockdev-utils-2.19-7.el8.x86_64", "Installed: mdadm-4.1-4.el8.x86_64", "Installed: nss-softokn-3.41.0-5.el8.x86_64", "Installed: libbytesize-1.4-1.el8.x86_64", "Installed: nss-softokn-freebl-3.41.0-5.el8.x86_64", "Installed: nss-sysinit-3.41.0-5.el8.x86_64", "Installed: nss-util-3.41.0-5.el8.x86_64", "Installed: volume_key-libs-0.3.11-4.el8.x86_64", "Installed: ndctl-62-2.el8.x86_64", "Installed: ndctl-libs-62-2.el8.x86_64", "Installed: blivet-data-1:3.1.0-9.el8.noarch", "Installed: python3-pyparted-1:3.11.0-13.el8.x86_64", "Installed: python3-blivet-1:3.1.0-9.el8.noarch", "Installed: python3-blockdev-2.19-7.el8.x86_64", "Installed: daxctl-libs-62-2.el8.x86_64", "Installed: gdisk-1.0.3-6.el8.x86_64", "Installed: python3-bytesize-1.4-1.el8.x86_64", "Installed: device-mapper-event-8:1.02.155-6.el8.x86_64", "Installed: device-mapper-event-libs-8:1.02.155-6.el8.x86_64", "Installed: libaio-0.3.110-12.el8.x86_64", "Installed: device-mapper-multipath-0.7.8-7.el8.x86_64", "Installed: device-mapper-multipath-libs-0.7.8-7.el8.x86_64", "Installed: device-mapper-persistent-data-0.7.6-1.el8.x86_64", "Installed: lsof-4.91-2.el8.x86_64", "Installed: libblockdev-2.19-7.el8.x86_64", "Installed: lvm2-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-crypto-2.19-7.el8.x86_64", "Installed: libblockdev-dm-2.19-7.el8.x86_64", "Installed: lvm2-libs-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-fs-2.19-7.el8.x86_64", "Installed: libblockdev-kbd-2.19-7.el8.x86_64", "Installed: libblockdev-loop-2.19-7.el8.x86_64", "Installed: libblockdev-lvm-2.19-7.el8.x86_64", "Installed: libblockdev-mdraid-2.19-7.el8.x86_64", "Installed: libblockdev-mpath-2.19-7.el8.x86_64", "Installed: nspr-4.20.0-2.el8.x86_64", "Installed: libblockdev-nvdimm-2.19-7.el8.x86_64"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": [], "changed": false, "leaves": [], "mounts": [], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [], "changed": false, "failed": false, "leaves": [], "mounts": [], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_filesystem_one_disk.yml:11 included: /tmp/tmpfvho35gc/tests/get_unused_disk.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Find unused disks in the system] ********************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:2 [WARNING]: The value 0 (type int) in a string field was converted to '0' (type string). If this does not look like what you expect, quote the entire value to ensure it does not change. ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "disks": ["vdb"]} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:8 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"unused_disks": ["vdb"]}, "changed": false} TASK [Exit playbook when there's no unused disks in the system] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:13 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} META: TASK [Print unused disks] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:19 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "unused_disks": [ "vdb" ] } TASK [Initialize a disk device with the default fs type] *************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_filesystem_one_disk.yml:15 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "msg": "Nothing to do", "rc": 0, "results": ["Installed: python3-blivet"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'name': 'test1', 'type': 'disk', 'mount_point': '/opt/test1', 'disks': ['vdb']}) => {"ansible_facts": {"volumes": [{"disks": ["vdb"], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "present", "type": "disk"}]}, "ansible_loop_var": "volume", "changed": false, "volume": {"disks": ["vdb"], "mount_point": "/opt/test1", "name": "test1", "type": "disk"}} TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": [{"disks": ["vdb"], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "present", "type": "disk"}]}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [ { "disks": [ "vdb" ], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "present", "type": "disk" } ] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": ["create format xfs on /dev/vdb"], "changed": true, "leaves": ["/dev/sr0", "/dev/vda1", "/dev/vdb", "/dev/vdc"], "mounts": [{"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/vdb", "state": "mounted"}], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [ "create format xfs on /dev/vdb" ], "changed": true, "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vdb", "/dev/vdc" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/vdb", "state": "mounted" } ], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'src': '/dev/vdb', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/vdb", "state": "mounted"}, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/vdb"} TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "cmd": ["systemctl", "daemon-reload"], "delta": "0:00:00.753172", "end": "2019-05-26 06:42:19.246052", "rc": 0, "start": "2019-05-26 06:42:18.492880", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_filesystem_one_disk.yml:25 included: /tmp/tmpfvho35gc/tests/verify-role-results.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:1 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"storage_tests_list": ["verify-volumes"]}, "changed": false} TASK [Print out pool information] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:4 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} TASK [Print out volume information] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:9 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "_storage_volumes_list": [ { "disks": [ "vdb" ], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "present", "type": "disk" } ] } TASK [Ensure the specified devices in the role were correctly configured in the user's system] ************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:14 included: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Read the /etc/fstab file for volume existence] ******************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "cmd": ["cat", "/etc/fstab"], "delta": "0:00:00.005632", "end": "2019-05-26 06:42:23.589156", "rc": 0, "start": "2019-05-26 06:42:23.583524", "stderr": "", "stderr_lines": [], "stdout": "\n#\n# /etc/fstab\n# Created by anaconda on Thu Apr 4 08:39:20 2019\n#\n# Accessible filesystems, by reference, are maintained under '/dev/disk/'.\n# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.\n#\n# After editing this file, run 'systemctl daemon-reload' to update systemd\n# units generated from this file.\n#\nUUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0\n/dev/vdb /opt/test1 xfs defaults 0 0", "stdout_lines": ["", "#", "# /etc/fstab", "# Created by anaconda on Thu Apr 4 08:39:20 2019", "#", "# Accessible filesystems, by reference, are maintained under '/dev/disk/'.", "# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.", "#", "# After editing this file, run 'systemctl daemon-reload' to update systemd", "# units generated from this file.", "#", "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0", "/dev/vdb /opt/test1 xfs defaults 0 0"]} TASK [Verify the volumes listed in storage_pools were correctly managed] *********************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:7 TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:14 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vdb) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vdb", "device": "/dev/vdb"} TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:21 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vda1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vda1", "device": "/dev/vda1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vdb) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vdb", "device": "/dev/vdb"} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:28 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_mount_devs": ["/dev/vda1", "/dev/vdb"], "blivet_mount_devs": ["/dev/vdb"], "removed_volumes": []}, "changed": false} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:33 TASK [Verify the volumes with no pool were correctly managed] ********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:40 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'disk', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'disks': ['vdb']}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "disks": [ "vdb" ], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "present", "type": "disk" }, "msg": "All assertions passed" } TASK [Verify the present volumes with no pool were correctly managed] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:52 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'disk', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'disks': ['vdb']}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "disks": [ "vdb" ], "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "size": 0, "state": "present", "type": "disk" }, "idx": 0, "msg": "All assertions passed" } TASK [Verify the absent volumes with no pool are not in ansible_mounts] ************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:65 META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=44 changed=5 unreachable=0 failed=0 skipped=23 rescued=0 ignored=0 + cd /tmp/tmpfvho35gc/tests; TEST_SUBJECTS=/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TEST_ARTIFACTS=/tmp/linux-system-role-test-work-pull-linux-system-roles_storage-25-db52957-rhel-8-df0ousve/artifacts ansible-playbook -vv --inventory=/usr/share/ansible/inventory/standard-inventory-qcow2 /tmp/tmpfvho35gc/_setup.yml /tmp/tmpfvho35gc/tests/tests_lvm_multiple_disks_multiple_volumes.yml ansible-playbook 2.8.0 config file = /home/till/.ansible.cfg configured module search path = ['/home/till/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 3.7.3 (default, May 11 2019, 00:38:04) [GCC 9.1.1 20190503 (Red Hat 9.1.1-1)] Using /home/till/.ansible.cfg as config file PLAYBOOK: _setup.yml *************************************************************************************************************************************************************************************** 2 plays in /tmp/tmpfvho35gc/_setup.yml PLAY [Fail when only localhost is available] *************************************************************************************************************************************************************** META: ran handlers TASK [debug] *********************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/_setup.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "groups": { "all": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "localhost": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "subjects": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "ungrouped": [] } } TASK [fail] ************************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:7 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} META: ran handlers META: ran handlers PLAY [Setup repos] ***************************************************************************************************************************************************************************************** META: ran handlers TASK [set up internal repositories] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:16 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true} META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=2 changed=1 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 PLAYBOOK: tests_lvm_multiple_disks_multiple_volumes.yml **************************************************************************************************************************************************** 1 plays in /tmp/tmpfvho35gc/tests/tests_lvm_multiple_disks_multiple_volumes.yml PLAY [all] ************************************************************************************************************************************************************************************************* TASK [Gathering Facts] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/tests_lvm_multiple_disks_multiple_volumes.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_all_ipv4_addresses": ["10.0.2.15"], "ansible_all_ipv6_addresses": ["fec0::5054:ff:fe12:3456", "fe80::5054:ff:fe12:3456"], "ansible_apparmor": {"status": "disabled"}, "ansible_architecture": "x86_64", "ansible_bios_date": "04/01/2014", "ansible_bios_version": "1.12.0-2.fc30", "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": "ttyS0,115200n8", "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_date_time": {"date": "2019-05-26", "day": "26", "epoch": "1558867453", "hour": "06", "iso8601": "2019-05-26T10:44:13Z", "iso8601_basic": "20190526T064413360540", "iso8601_basic_short": "20190526T064413", "iso8601_micro": "2019-05-26T10:44:13.360685Z", "minute": "44", "month": "05", "second": "13", "time": "06:44:13", "tz": "EDT", "tz_offset": "-0400", "weekday": "Sonntag", "weekday_number": "0", "weeknumber": "20", "year": "2019"}, "ansible_default_ipv4": {"address": "10.0.2.15", "alias": "eth0", "broadcast": "10.0.2.255", "gateway": "10.0.2.2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "netmask": "255.255.255.0", "network": "10.0.2.0", "type": "ether"}, "ansible_default_ipv6": {"address": "fec0::5054:ff:fe12:3456", "gateway": "fe80::2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "prefix": "64", "scope": "site", "type": "ether"}, "ansible_device_links": {"ids": {"sr0": ["ata-QEMU_DVD-ROM_QM00003"]}, "labels": {"sr0": ["cidata"]}, "masters": {}, "uuids": {"sr0": ["2019-05-26-12-42-29-00"], "vda1": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}}, "ansible_devices": {"sr0": {"holders": [], "host": "IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]", "links": {"ids": ["ata-QEMU_DVD-ROM_QM00003"], "labels": ["cidata"], "masters": [], "uuids": ["2019-05-26-12-42-29-00"]}, "model": "QEMU DVD-ROM", "partitions": {}, "removable": "1", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "728", "sectorsize": "2048", "size": "364.00 KB", "support_discard": "0", "vendor": "QEMU", "virtual": 1}, "vda": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {"vda1": {"holders": [], "links": {"ids": [], "labels": [], "masters": [], "uuids": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}, "sectors": "20969439", "sectorsize": 512, "size": "10.00 GB", "start": "2048", "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdb": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdc": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}}, "ansible_distribution": "RedHat", "ansible_distribution_file_parsed": true, "ansible_distribution_file_path": "/etc/redhat-release", "ansible_distribution_file_search_string": "Red Hat", "ansible_distribution_file_variety": "RedHat", "ansible_distribution_major_version": "8", "ansible_distribution_release": "Ootpa", "ansible_distribution_version": "8.0", "ansible_dns": {"nameservers": ["10.0.2.3"], "search": ["virt.pnr.lab.eng.rdu2.redhat.com"]}, "ansible_domain": "virt.pnr.lab.eng.rdu2.redhat.com", "ansible_effective_group_id": 0, "ansible_effective_user_id": 0, "ansible_env": {"DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "HOME": "/root", "LANG": "de_DE.UTF-8", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "LOGNAME": "root", "LS_COLORS": "rs=0:di=38;5;33:ln=38;5;51:mh=00:pi=40;38;5;11:so=38;5;13:do=38;5;5:bd=48;5;232;38;5;11:cd=48;5;232;38;5;3:or=48;5;232;38;5;9:mi=01;05;37;41:su=48;5;196;38;5;15:sg=48;5;11;38;5;16:ca=48;5;196;38;5;226:tw=48;5;10;38;5;16:ow=48;5;10;38;5;21:st=48;5;21;38;5;15:ex=38;5;40:*.tar=38;5;9:*.tgz=38;5;9:*.arc=38;5;9:*.arj=38;5;9:*.taz=38;5;9:*.lha=38;5;9:*.lz4=38;5;9:*.lzh=38;5;9:*.lzma=38;5;9:*.tlz=38;5;9:*.txz=38;5;9:*.tzo=38;5;9:*.t7z=38;5;9:*.zip=38;5;9:*.z=38;5;9:*.dz=38;5;9:*.gz=38;5;9:*.lrz=38;5;9:*.lz=38;5;9:*.lzo=38;5;9:*.xz=38;5;9:*.zst=38;5;9:*.tzst=38;5;9:*.bz2=38;5;9:*.bz=38;5;9:*.tbz=38;5;9:*.tbz2=38;5;9:*.tz=38;5;9:*.deb=38;5;9:*.rpm=38;5;9:*.jar=38;5;9:*.war=38;5;9:*.ear=38;5;9:*.sar=38;5;9:*.rar=38;5;9:*.alz=38;5;9:*.ace=38;5;9:*.zoo=38;5;9:*.cpio=38;5;9:*.7z=38;5;9:*.rz=38;5;9:*.cab=38;5;9:*.wim=38;5;9:*.swm=38;5;9:*.dwm=38;5;9:*.esd=38;5;9:*.jpg=38;5;13:*.jpeg=38;5;13:*.mjpg=38;5;13:*.mjpeg=38;5;13:*.gif=38;5;13:*.bmp=38;5;13:*.pbm=38;5;13:*.pgm=38;5;13:*.ppm=38;5;13:*.tga=38;5;13:*.xbm=38;5;13:*.xpm=38;5;13:*.tif=38;5;13:*.tiff=38;5;13:*.png=38;5;13:*.svg=38;5;13:*.svgz=38;5;13:*.mng=38;5;13:*.pcx=38;5;13:*.mov=38;5;13:*.mpg=38;5;13:*.mpeg=38;5;13:*.m2v=38;5;13:*.mkv=38;5;13:*.webm=38;5;13:*.ogm=38;5;13:*.mp4=38;5;13:*.m4v=38;5;13:*.mp4v=38;5;13:*.vob=38;5;13:*.qt=38;5;13:*.nuv=38;5;13:*.wmv=38;5;13:*.asf=38;5;13:*.rm=38;5;13:*.rmvb=38;5;13:*.flc=38;5;13:*.avi=38;5;13:*.fli=38;5;13:*.flv=38;5;13:*.gl=38;5;13:*.dl=38;5;13:*.xcf=38;5;13:*.xwd=38;5;13:*.yuv=38;5;13:*.cgm=38;5;13:*.emf=38;5;13:*.ogv=38;5;13:*.ogx=38;5;13:*.aac=38;5;45:*.au=38;5;45:*.flac=38;5;45:*.m4a=38;5;45:*.mid=38;5;45:*.midi=38;5;45:*.mka=38;5;45:*.mp3=38;5;45:*.mpc=38;5;45:*.ogg=38;5;45:*.ra=38;5;45:*.wav=38;5;45:*.oga=38;5;45:*.opus=38;5;45:*.spx=38;5;45:*.xspf=38;5;45:", "MAIL": "/var/mail/root", "PATH": "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "PWD": "/root", "SELINUX_LEVEL_REQUESTED": "", "SELINUX_ROLE_REQUESTED": "", "SELINUX_USE_CURRENT_RANGE": "", "SHELL": "/bin/bash", "SHLVL": "2", "SSH_CLIENT": "10.0.2.2 38710 22", "SSH_CONNECTION": "10.0.2.2 38710 10.0.2.15 22", "SSH_TTY": "/dev/pts/0", "TERM": "xterm-256color", "USER": "root", "XDG_RUNTIME_DIR": "/run/user/0", "XDG_SESSION_ID": "1", "_": "/usr/libexec/platform-python"}, "ansible_eth0": {"active": true, "device": "eth0", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "off [requested on]", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "off [fixed]", "netns_local": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "off [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "on [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "off", "tcp_segmentation_offload": "off", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "tx_checksumming": "off", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "off [fixed]", "tx_nocache_copy": "off", "tx_scatter_gather": "off [fixed]", "tx_scatter_gather_fraglist": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_tcp6_segmentation": "off [fixed]", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off [fixed]", "tx_tcp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "off [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "10.0.2.15", "broadcast": "10.0.2.255", "netmask": "255.255.255.0", "network": "10.0.2.0"}, "ipv6": [{"address": "fec0::5054:ff:fe12:3456", "prefix": "64", "scope": "site"}, {"address": "fe80::5054:ff:fe12:3456", "prefix": "64", "scope": "link"}], "macaddress": "52:54:00:12:34:56", "module": "virtio_net", "mtu": 1500, "pciid": "virtio0", "promisc": false, "speed": -1, "timestamping": ["tx_software", "rx_software", "software"], "type": "ether"}, "ansible_fibre_channel_wwn": [], "ansible_fips": false, "ansible_form_factor": "Other", "ansible_fqdn": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_hostname": "ibm-p8-kvm-03-guest-02", "ansible_hostnqn": "", "ansible_interfaces": ["eth0", "lo"], "ansible_is_chroot": false, "ansible_iscsi_iqn": "", "ansible_kernel": "4.18.0-80.el8.x86_64", "ansible_lo": {"active": true, "device": "lo", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "on", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "on [fixed]", "netns_local": "on [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "on [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "off [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "on", "tcp_segmentation_offload": "on", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "tx_checksumming": "on", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "on [fixed]", "tx_nocache_copy": "off [fixed]", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tx_sctp_segmentation": "on", "tx_tcp6_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp_segmentation": "on", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "on [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "127.0.0.1", "broadcast": "host", "netmask": "255.0.0.0", "network": "127.0.0.0"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "mtu": 65536, "promisc": false, "timestamping": ["tx_software", "rx_software", "software"], "type": "loopback"}, "ansible_local": {}, "ansible_lsb": {}, "ansible_machine": "x86_64", "ansible_machine_id": "82f3b023080543488647d22e2c0adee6", "ansible_memfree_mb": 387, "ansible_memory_mb": {"nocache": {"free": 625, "used": 195}, "real": {"free": 387, "total": 820, "used": 433}, "swap": {"cached": 0, "free": 0, "total": 0, "used": 0}}, "ansible_memtotal_mb": 820, "ansible_mounts": [{"block_available": 2290090, "block_size": 4096, "block_total": 2618619, "block_used": 328529, "device": "/dev/vda1", "fstype": "xfs", "inode_available": 5207720, "inode_total": 5242352, "inode_used": 34632, "mount": "/", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 9380208640, "size_total": 10725863424, "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}], "ansible_nodename": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_os_family": "RedHat", "ansible_pkg_mgr": "dnf", "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": ["ttyS0", "ttyS0,115200n8"], "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_processor": ["0", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "1", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "2", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz"], "ansible_processor_cores": 1, "ansible_processor_count": 3, "ansible_processor_threads_per_core": 1, "ansible_processor_vcpus": 3, "ansible_product_name": "Standard PC (i440FX + PIIX, 1996)", "ansible_product_serial": "NA", "ansible_product_uuid": "NA", "ansible_product_version": "pc-i440fx-3.1", "ansible_python": {"executable": "/usr/libexec/platform-python", "has_sslcontext": true, "type": "cpython", "version": {"major": 3, "micro": 8, "minor": 6, "releaselevel": "final", "serial": 0}, "version_info": [3, 6, 8, "final", 0]}, "ansible_python_version": "3.6.8", "ansible_real_group_id": 0, "ansible_real_user_id": 0, "ansible_selinux": {"config_mode": "enforcing", "mode": "enforcing", "policyvers": 31, "status": "enabled", "type": "targeted"}, "ansible_selinux_python_present": true, "ansible_service_mgr": "systemd", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMWD6ZH4Q5qRaqQMk5uPNhyRZHYwohzW3LQyQ1GV1nvAVVqA/8zLDLXZgJhbS9HVMnUPlHFvxV79KTGlOUlL+ko=", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIJRmaImO3tC6zXfHOTeKchUAvpnvsPJ7uHi2V1Ym2ucQ", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABAQC7EpNfwdANrPWNmSaHvVs6ZRR/7SvgAlveClp//b12/9mj58mcFbTMX9AmPftiaG2lK5BTR/++3M/hyzKSSRiCQ0fujMYx1AMlOnm/nb5hKLDU61lKjDTzSc8PI7SmH+2Z22O1elX605jCJdIQqLb23VRUbwG3WK8q2BMwvait8jpA5LEEaTNqnImfVlzN3j6o2fNRhMEIpbQNDvFPTX5n1v1d5nWz2hpmbCyvedI8tvaakZwn5PUMJEydX+tcAldGUmFrP5pwk/ptqt1pgPH3jaGCFI7Fl8zCj+iYyvKeaeuApnpTZsyaBGIhczBj18WzRSdZ/M7DaCDgDsMAofLD", "ansible_swapfree_mb": 0, "ansible_swaptotal_mb": 0, "ansible_system": "Linux", "ansible_system_capabilities": ["cap_chown", "cap_dac_override", "cap_dac_read_search", "cap_fowner", "cap_fsetid", "cap_kill", "cap_setgid", "cap_setuid", "cap_setpcap", "cap_linux_immutable", "cap_net_bind_service", "cap_net_broadcast", "cap_net_admin", "cap_net_raw", "cap_ipc_lock", "cap_ipc_owner", "cap_sys_module", "cap_sys_rawio", "cap_sys_chroot", "cap_sys_ptrace", "cap_sys_pacct", "cap_sys_admin", "cap_sys_boot", "cap_sys_nice", "cap_sys_resource", "cap_sys_time", "cap_sys_tty_config", "cap_mknod", "cap_lease", "cap_audit_write", "cap_audit_control", "cap_setfcap", "cap_mac_override", "cap_mac_admin", "cap_syslog", "cap_wake_alarm", "cap_block_suspend", "cap_audit_read+ep"], "ansible_system_capabilities_enforced": "True", "ansible_system_vendor": "QEMU", "ansible_uptime_seconds": 99, "ansible_user_dir": "/root", "ansible_user_gecos": "root", "ansible_user_gid": 0, "ansible_user_id": "root", "ansible_user_shell": "/bin/bash", "ansible_user_uid": 0, "ansible_userspace_architecture": "x86_64", "ansible_userspace_bits": "64", "ansible_virtualization_role": "guest", "ansible_virtualization_type": "kvm", "gather_subset": ["all"], "module_setup": true}, "changed": false} META: ran handlers TASK [include_role : storage] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_lvm_multiple_disks_multiple_volumes.yml:11 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "msg": "", "rc": 0, "results": ["Installed: python3-blivet", "Installed: libblockdev-part-2.19-7.el8.x86_64", "Installed: nss-3.41.0-5.el8.x86_64", "Installed: libblockdev-swap-2.19-7.el8.x86_64", "Installed: userspace-rcu-0.10.1-2.el8.x86_64", "Installed: libblockdev-utils-2.19-7.el8.x86_64", "Installed: mdadm-4.1-4.el8.x86_64", "Installed: nss-softokn-3.41.0-5.el8.x86_64", "Installed: libbytesize-1.4-1.el8.x86_64", "Installed: nss-softokn-freebl-3.41.0-5.el8.x86_64", "Installed: nss-sysinit-3.41.0-5.el8.x86_64", "Installed: nss-util-3.41.0-5.el8.x86_64", "Installed: volume_key-libs-0.3.11-4.el8.x86_64", "Installed: ndctl-62-2.el8.x86_64", "Installed: ndctl-libs-62-2.el8.x86_64", "Installed: blivet-data-1:3.1.0-9.el8.noarch", "Installed: python3-pyparted-1:3.11.0-13.el8.x86_64", "Installed: python3-blivet-1:3.1.0-9.el8.noarch", "Installed: python3-blockdev-2.19-7.el8.x86_64", "Installed: daxctl-libs-62-2.el8.x86_64", "Installed: gdisk-1.0.3-6.el8.x86_64", "Installed: python3-bytesize-1.4-1.el8.x86_64", "Installed: device-mapper-event-8:1.02.155-6.el8.x86_64", "Installed: device-mapper-event-libs-8:1.02.155-6.el8.x86_64", "Installed: libaio-0.3.110-12.el8.x86_64", "Installed: device-mapper-multipath-0.7.8-7.el8.x86_64", "Installed: device-mapper-multipath-libs-0.7.8-7.el8.x86_64", "Installed: device-mapper-persistent-data-0.7.6-1.el8.x86_64", "Installed: lsof-4.91-2.el8.x86_64", "Installed: libblockdev-2.19-7.el8.x86_64", "Installed: lvm2-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-crypto-2.19-7.el8.x86_64", "Installed: libblockdev-dm-2.19-7.el8.x86_64", "Installed: lvm2-libs-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-fs-2.19-7.el8.x86_64", "Installed: libblockdev-kbd-2.19-7.el8.x86_64", "Installed: libblockdev-loop-2.19-7.el8.x86_64", "Installed: libblockdev-lvm-2.19-7.el8.x86_64", "Installed: libblockdev-mdraid-2.19-7.el8.x86_64", "Installed: libblockdev-mpath-2.19-7.el8.x86_64", "Installed: nspr-4.20.0-2.el8.x86_64", "Installed: libblockdev-nvdimm-2.19-7.el8.x86_64"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": [], "changed": false, "leaves": [], "mounts": [], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [], "changed": false, "failed": false, "leaves": [], "mounts": [], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_lvm_multiple_disks_multiple_volumes.yml:14 included: /tmp/tmpfvho35gc/tests/get_unused_disk.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Find unused disks in the system] ********************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "disks": ["vdb", "vdc"]} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:8 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"unused_disks": ["vdb", "vdc"]}, "changed": false} TASK [Exit playbook when there's no unused disks in the system] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:13 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} META: TASK [Print unused disks] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:19 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "unused_disks": [ "vdb", "vdc" ] } TASK [debug] *********************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_lvm_multiple_disks_multiple_volumes.yml:20 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} META: TASK [Create a logical volume spanning two physical volumes that changes its mount location] *************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_lvm_multiple_disks_multiple_volumes.yml:26 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "msg": "Nothing to do", "rc": 0, "results": ["Installed: python3-blivet"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'name': 'phi', 'disks': ['vdb', 'vdc'], 'volumes': [{'name': 'test1', 'size': '4g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb", "vdc"], "name": "phi", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "4g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}]}, "ansible_loop_var": "pool", "changed": false, "pool": {"disks": ["vdb", "vdc"], "name": "phi", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "4g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'state': 'present', 'type': 'lvm', 'name': 'phi', 'disks': ['vdb', 'vdc'], 'volumes': [{'name': 'test1', 'size': '4g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]}, {'name': 'test1', 'size': '4g', 'mount_point': '/opt/test1'}]) => {"ansible_facts": {"vol_defaults": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}], "vol_pools": ["phi"], "vols": [{"mount_point": "/opt/test1", "name": "test1", "size": "4g"}]}, "ansible_loop_var": "item", "changed": false, "item": [{"disks": ["vdb", "vdc"], "name": "phi", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "4g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}, {"mount_point": "/opt/test1", "name": "test1", "size": "4g"}]} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'state': 'present', 'type': 'lvm', 'name': 'phi', 'disks': ['vdb', 'vdc'], 'volumes': [{'name': 'test1', 'size': '4g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]) => {"ansible_facts": {"vol_defaults": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}], "vol_pools": ["phi", "phi"], "vols": [{"mount_point": "/opt/test1", "name": "test1", "size": "4g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}, "ansible_loop_var": "item", "changed": false, "item": [{"disks": ["vdb", "vdc"], "name": "phi", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "4g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'name': 'test1', 'size': '4g', 'mount_point': '/opt/test1'}, {'state': 'present', 'type': 'lvm', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid'}]) => {"ansible_facts": {"all_vols": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "phi", "size": "4g", "state": "present", "type": "lvm"}]}, "ansible_loop_var": "item", "changed": false, "idx": 0, "item": [{"mount_point": "/opt/test1", "name": "test1", "size": "4g"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}]} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}, {'state': 'present', 'type': 'lvm', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid'}]) => {"ansible_facts": {"all_vols": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "phi", "size": "4g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "phi", "size": "4g", "state": "present", "type": "lvm"}]}, "ansible_loop_var": "item", "changed": false, "idx": 1, "item": [{"mount_point": "/opt/test2", "name": "test2", "size": "4g"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'phi', 'disks': ['vdb', 'vdc'], 'volumes': [{'name': 'test1', 'size': '4g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]}) => {"ansible_facts": {"vols_by_pool": {"phi": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "phi", "size": "4g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "phi", "size": "4g", "state": "present", "type": "lvm"}]}}, "ansible_loop_var": "item", "changed": false, "item": {"disks": ["vdb", "vdc"], "name": "phi", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "4g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'phi', 'disks': ['vdb', 'vdc'], 'volumes': [{'name': 'test1', 'size': '4g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '4g', 'mount_point': '/opt/test2'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb", "vdc"], "name": "phi", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "phi", "size": "4g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "phi", "size": "4g", "state": "present", "type": "lvm"}]}]}, "ansible_loop_var": "pool", "changed": false, "idx": 0, "pool": {"disks": ["vdb", "vdc"], "name": "phi", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "4g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "4g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": [{"disks": ["vdb", "vdc"], "name": "phi", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "phi", "size": "4g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "phi", "size": "4g", "state": "present", "type": "lvm"}]}]}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [ { "disks": [ "vdb", "vdc" ], "name": "phi", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "phi", "size": "4g", "state": "present", "type": "lvm" }, { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "phi", "size": "4g", "state": "present", "type": "lvm" } ] } ] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": ["create format lvmpv on /dev/vdc", "create format lvmpv on /dev/vdb", "create device /dev/phi", "create device /dev/mapper/phi-test2", "create format xfs on /dev/mapper/phi-test2", "create device /dev/mapper/phi-test1", "create format xfs on /dev/mapper/phi-test1"], "changed": true, "leaves": ["/dev/sr0", "/dev/vda1", "/dev/mapper/phi-test1", "/dev/mapper/phi-test2"], "mounts": [{"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/phi-test1", "state": "mounted"}, {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/phi-test2", "state": "mounted"}], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [ "create format lvmpv on /dev/vdc", "create format lvmpv on /dev/vdb", "create device /dev/phi", "create device /dev/mapper/phi-test2", "create format xfs on /dev/mapper/phi-test2", "create device /dev/mapper/phi-test1", "create format xfs on /dev/mapper/phi-test1" ], "changed": true, "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/mapper/phi-test1", "/dev/mapper/phi-test2" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/phi-test1", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/phi-test2", "state": "mounted" } ], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'src': '/dev/mapper/phi-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/phi-test1", "state": "mounted"}, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/phi-test1"} changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'src': '/dev/mapper/phi-test2', 'path': '/opt/test2', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/phi-test2", "state": "mounted"}, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/phi-test2"} TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "cmd": ["systemctl", "daemon-reload"], "delta": "0:00:00.683444", "end": "2019-05-26 06:45:24.761710", "rc": 0, "start": "2019-05-26 06:45:24.078266", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_lvm_multiple_disks_multiple_volumes.yml:41 included: /tmp/tmpfvho35gc/tests/verify-role-results.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:1 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"storage_tests_list": ["verify-volumes"]}, "changed": false} TASK [Print out pool information] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:4 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "_storage_pools_list": [ { "disks": [ "vdb", "vdc" ], "name": "phi", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "phi", "size": "4g", "state": "present", "type": "lvm" }, { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "phi", "size": "4g", "state": "present", "type": "lvm" } ] } ] } TASK [Print out volume information] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:9 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} TASK [Ensure the specified devices in the role were correctly configured in the user's system] ************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:14 included: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Read the /etc/fstab file for volume existence] ******************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "cmd": ["cat", "/etc/fstab"], "delta": "0:00:00.018849", "end": "2019-05-26 06:45:29.378470", "rc": 0, "start": "2019-05-26 06:45:29.359621", "stderr": "", "stderr_lines": [], "stdout": "\n#\n# /etc/fstab\n# Created by anaconda on Thu Apr 4 08:39:20 2019\n#\n# Accessible filesystems, by reference, are maintained under '/dev/disk/'.\n# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.\n#\n# After editing this file, run 'systemctl daemon-reload' to update systemd\n# units generated from this file.\n#\nUUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0\n/dev/mapper/phi-test1 /opt/test1 xfs defaults 0 0\n/dev/mapper/phi-test2 /opt/test2 xfs defaults 0 0", "stdout_lines": ["", "#", "# /etc/fstab", "# Created by anaconda on Thu Apr 4 08:39:20 2019", "#", "# Accessible filesystems, by reference, are maintained under '/dev/disk/'.", "# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.", "#", "# After editing this file, run 'systemctl daemon-reload' to update systemd", "# units generated from this file.", "#", "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0", "/dev/mapper/phi-test1 /opt/test1 xfs defaults 0 0", "/dev/mapper/phi-test2 /opt/test2 xfs defaults 0 0"]} TASK [Verify the volumes listed in storage_pools were correctly managed] *********************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:7 included: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"mount_entries": []}, "changed": false} TASK [gather mounts related to each volume in the pool] **************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'phi'}) => {"ansible_facts": {"mount_entries": [[{"block_available": 1030436, "block_size": 4096, "block_total": 1046016, "block_used": 15580, "device": "/dev/mapper/phi-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4220665856, "size_total": 4284481536, "uuid": "5b830e9c-4726-47a7-9b01-6e89f2e68599"}]]}, "ansible_loop_var": "current_volume", "changed": false, "current_volume": {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "phi", "size": "4g", "state": "present", "type": "lvm"}} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test2', 'pool': 'phi'}) => {"ansible_facts": {"mount_entries": [[{"block_available": 1030436, "block_size": 4096, "block_total": 1046016, "block_used": 15580, "device": "/dev/mapper/phi-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4220665856, "size_total": 4284481536, "uuid": "5b830e9c-4726-47a7-9b01-6e89f2e68599"}], [{"block_available": 1030436, "block_size": 4096, "block_total": 1046016, "block_used": 15580, "device": "/dev/mapper/phi-test2", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4220665856, "size_total": 4284481536, "uuid": "b15ef1b9-ba9d-48fb-8465-32af96decb4d"}]]}, "ansible_loop_var": "current_volume", "changed": false, "current_volume": {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "phi", "size": "4g", "state": "present", "type": "lvm"}} TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:12 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'phi'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "phi", "size": "4g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test2', 'pool': 'phi'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "phi", "size": "4g", "state": "present", "type": "lvm" }, "idx": 1, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:25 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'phi'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "phi", "size": "4g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test2', 'pool': 'phi'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "phi", "size": "4g", "state": "present", "type": "lvm" }, "idx": 1, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab.] **************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:38 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'phi'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "phi", "size": "4g", "state": "present", "type": "lvm" }, "msg": "All assertions passed" } ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test2', 'pool': 'phi'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "phi", "size": "4g", "state": "present", "type": "lvm" }, "msg": "All assertions passed" } TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:14 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/phi-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/phi-test1", "device": "/dev/mapper/phi-test1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/phi-test2) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/phi-test2", "device": "/dev/mapper/phi-test2"} TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:21 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vda1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vda1", "device": "/dev/vda1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/phi-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/phi-test1", "device": "/dev/mapper/phi-test1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/phi-test2) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/phi-test2", "device": "/dev/mapper/phi-test2"} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:28 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_mount_devs": ["/dev/vda1", "/dev/mapper/phi-test1", "/dev/mapper/phi-test2"], "blivet_mount_devs": ["/dev/mapper/phi-test1", "/dev/mapper/phi-test2"], "removed_volumes": []}, "changed": false} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:33 TASK [Verify the volumes with no pool were correctly managed] ********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:40 TASK [Verify the present volumes with no pool were correctly managed] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:52 TASK [Verify the absent volumes with no pool are not in ansible_mounts] ************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:65 META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=52 changed=5 unreachable=0 failed=0 skipped=21 rescued=0 ignored=0 + cd /tmp/tmpfvho35gc/tests; TEST_SUBJECTS=/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TEST_ARTIFACTS=/tmp/linux-system-role-test-work-pull-linux-system-roles_storage-25-db52957-rhel-8-df0ousve/artifacts ansible-playbook -vv --inventory=/usr/share/ansible/inventory/standard-inventory-qcow2 /tmp/tmpfvho35gc/_setup.yml /tmp/tmpfvho35gc/tests/tests_lvm_one_disk_multiple_volumes.yml ansible-playbook 2.8.0 config file = /home/till/.ansible.cfg configured module search path = ['/home/till/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 3.7.3 (default, May 11 2019, 00:38:04) [GCC 9.1.1 20190503 (Red Hat 9.1.1-1)] Using /home/till/.ansible.cfg as config file PLAYBOOK: _setup.yml *************************************************************************************************************************************************************************************** 2 plays in /tmp/tmpfvho35gc/_setup.yml PLAY [Fail when only localhost is available] *************************************************************************************************************************************************************** META: ran handlers TASK [debug] *********************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/_setup.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "groups": { "all": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "localhost": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "subjects": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "ungrouped": [] } } TASK [fail] ************************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:7 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} META: ran handlers META: ran handlers PLAY [Setup repos] ***************************************************************************************************************************************************************************************** META: ran handlers TASK [set up internal repositories] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:16 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true} META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=2 changed=1 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 PLAYBOOK: tests_lvm_one_disk_multiple_volumes.yml ********************************************************************************************************************************************************** 1 plays in /tmp/tmpfvho35gc/tests/tests_lvm_one_disk_multiple_volumes.yml PLAY [all] ************************************************************************************************************************************************************************************************* TASK [Gathering Facts] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/tests_lvm_one_disk_multiple_volumes.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_all_ipv4_addresses": ["10.0.2.15"], "ansible_all_ipv6_addresses": ["fec0::5054:ff:fe12:3456", "fe80::5054:ff:fe12:3456"], "ansible_apparmor": {"status": "disabled"}, "ansible_architecture": "x86_64", "ansible_bios_date": "04/01/2014", "ansible_bios_version": "1.12.0-2.fc30", "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": "ttyS0,115200n8", "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_date_time": {"date": "2019-05-26", "day": "26", "epoch": "1558867642", "hour": "06", "iso8601": "2019-05-26T10:47:22Z", "iso8601_basic": "20190526T064722417288", "iso8601_basic_short": "20190526T064722", "iso8601_micro": "2019-05-26T10:47:22.417392Z", "minute": "47", "month": "05", "second": "22", "time": "06:47:22", "tz": "EDT", "tz_offset": "-0400", "weekday": "Sonntag", "weekday_number": "0", "weeknumber": "20", "year": "2019"}, "ansible_default_ipv4": {"address": "10.0.2.15", "alias": "eth0", "broadcast": "10.0.2.255", "gateway": "10.0.2.2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "netmask": "255.255.255.0", "network": "10.0.2.0", "type": "ether"}, "ansible_default_ipv6": {"address": "fec0::5054:ff:fe12:3456", "gateway": "fe80::2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "prefix": "64", "scope": "site", "type": "ether"}, "ansible_device_links": {"ids": {"sr0": ["ata-QEMU_DVD-ROM_QM00003"]}, "labels": {"sr0": ["cidata"]}, "masters": {}, "uuids": {"sr0": ["2019-05-26-12-45-39-00"], "vda1": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}}, "ansible_devices": {"sr0": {"holders": [], "host": "IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]", "links": {"ids": ["ata-QEMU_DVD-ROM_QM00003"], "labels": ["cidata"], "masters": [], "uuids": ["2019-05-26-12-45-39-00"]}, "model": "QEMU DVD-ROM", "partitions": {}, "removable": "1", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "728", "sectorsize": "2048", "size": "364.00 KB", "support_discard": "0", "vendor": "QEMU", "virtual": 1}, "vda": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {"vda1": {"holders": [], "links": {"ids": [], "labels": [], "masters": [], "uuids": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}, "sectors": "20969439", "sectorsize": 512, "size": "10.00 GB", "start": "2048", "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdb": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdc": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}}, "ansible_distribution": "RedHat", "ansible_distribution_file_parsed": true, "ansible_distribution_file_path": "/etc/redhat-release", "ansible_distribution_file_search_string": "Red Hat", "ansible_distribution_file_variety": "RedHat", "ansible_distribution_major_version": "8", "ansible_distribution_release": "Ootpa", "ansible_distribution_version": "8.0", "ansible_dns": {"nameservers": ["10.0.2.3"], "search": ["virt.pnr.lab.eng.rdu2.redhat.com"]}, "ansible_domain": "virt.pnr.lab.eng.rdu2.redhat.com", "ansible_effective_group_id": 0, "ansible_effective_user_id": 0, "ansible_env": {"DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "HOME": "/root", "LANG": "de_DE.UTF-8", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "LOGNAME": "root", "LS_COLORS": "rs=0:di=38;5;33:ln=38;5;51:mh=00:pi=40;38;5;11:so=38;5;13:do=38;5;5:bd=48;5;232;38;5;11:cd=48;5;232;38;5;3:or=48;5;232;38;5;9:mi=01;05;37;41:su=48;5;196;38;5;15:sg=48;5;11;38;5;16:ca=48;5;196;38;5;226:tw=48;5;10;38;5;16:ow=48;5;10;38;5;21:st=48;5;21;38;5;15:ex=38;5;40:*.tar=38;5;9:*.tgz=38;5;9:*.arc=38;5;9:*.arj=38;5;9:*.taz=38;5;9:*.lha=38;5;9:*.lz4=38;5;9:*.lzh=38;5;9:*.lzma=38;5;9:*.tlz=38;5;9:*.txz=38;5;9:*.tzo=38;5;9:*.t7z=38;5;9:*.zip=38;5;9:*.z=38;5;9:*.dz=38;5;9:*.gz=38;5;9:*.lrz=38;5;9:*.lz=38;5;9:*.lzo=38;5;9:*.xz=38;5;9:*.zst=38;5;9:*.tzst=38;5;9:*.bz2=38;5;9:*.bz=38;5;9:*.tbz=38;5;9:*.tbz2=38;5;9:*.tz=38;5;9:*.deb=38;5;9:*.rpm=38;5;9:*.jar=38;5;9:*.war=38;5;9:*.ear=38;5;9:*.sar=38;5;9:*.rar=38;5;9:*.alz=38;5;9:*.ace=38;5;9:*.zoo=38;5;9:*.cpio=38;5;9:*.7z=38;5;9:*.rz=38;5;9:*.cab=38;5;9:*.wim=38;5;9:*.swm=38;5;9:*.dwm=38;5;9:*.esd=38;5;9:*.jpg=38;5;13:*.jpeg=38;5;13:*.mjpg=38;5;13:*.mjpeg=38;5;13:*.gif=38;5;13:*.bmp=38;5;13:*.pbm=38;5;13:*.pgm=38;5;13:*.ppm=38;5;13:*.tga=38;5;13:*.xbm=38;5;13:*.xpm=38;5;13:*.tif=38;5;13:*.tiff=38;5;13:*.png=38;5;13:*.svg=38;5;13:*.svgz=38;5;13:*.mng=38;5;13:*.pcx=38;5;13:*.mov=38;5;13:*.mpg=38;5;13:*.mpeg=38;5;13:*.m2v=38;5;13:*.mkv=38;5;13:*.webm=38;5;13:*.ogm=38;5;13:*.mp4=38;5;13:*.m4v=38;5;13:*.mp4v=38;5;13:*.vob=38;5;13:*.qt=38;5;13:*.nuv=38;5;13:*.wmv=38;5;13:*.asf=38;5;13:*.rm=38;5;13:*.rmvb=38;5;13:*.flc=38;5;13:*.avi=38;5;13:*.fli=38;5;13:*.flv=38;5;13:*.gl=38;5;13:*.dl=38;5;13:*.xcf=38;5;13:*.xwd=38;5;13:*.yuv=38;5;13:*.cgm=38;5;13:*.emf=38;5;13:*.ogv=38;5;13:*.ogx=38;5;13:*.aac=38;5;45:*.au=38;5;45:*.flac=38;5;45:*.m4a=38;5;45:*.mid=38;5;45:*.midi=38;5;45:*.mka=38;5;45:*.mp3=38;5;45:*.mpc=38;5;45:*.ogg=38;5;45:*.ra=38;5;45:*.wav=38;5;45:*.oga=38;5;45:*.opus=38;5;45:*.spx=38;5;45:*.xspf=38;5;45:", "MAIL": "/var/mail/root", "PATH": "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "PWD": "/root", "SELINUX_LEVEL_REQUESTED": "", "SELINUX_ROLE_REQUESTED": "", "SELINUX_USE_CURRENT_RANGE": "", "SHELL": "/bin/bash", "SHLVL": "2", "SSH_CLIENT": "10.0.2.2 40978 22", "SSH_CONNECTION": "10.0.2.2 40978 10.0.2.15 22", "SSH_TTY": "/dev/pts/0", "TERM": "xterm-256color", "USER": "root", "XDG_RUNTIME_DIR": "/run/user/0", "XDG_SESSION_ID": "1", "_": "/usr/libexec/platform-python"}, "ansible_eth0": {"active": true, "device": "eth0", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "off [requested on]", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "off [fixed]", "netns_local": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "off [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "on [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "off", "tcp_segmentation_offload": "off", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "tx_checksumming": "off", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "off [fixed]", "tx_nocache_copy": "off", "tx_scatter_gather": "off [fixed]", "tx_scatter_gather_fraglist": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_tcp6_segmentation": "off [fixed]", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off [fixed]", "tx_tcp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "off [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "10.0.2.15", "broadcast": "10.0.2.255", "netmask": "255.255.255.0", "network": "10.0.2.0"}, "ipv6": [{"address": "fec0::5054:ff:fe12:3456", "prefix": "64", "scope": "site"}, {"address": "fe80::5054:ff:fe12:3456", "prefix": "64", "scope": "link"}], "macaddress": "52:54:00:12:34:56", "module": "virtio_net", "mtu": 1500, "pciid": "virtio0", "promisc": false, "speed": -1, "timestamping": ["tx_software", "rx_software", "software"], "type": "ether"}, "ansible_fibre_channel_wwn": [], "ansible_fips": false, "ansible_form_factor": "Other", "ansible_fqdn": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_hostname": "ibm-p8-kvm-03-guest-02", "ansible_hostnqn": "", "ansible_interfaces": ["lo", "eth0"], "ansible_is_chroot": false, "ansible_iscsi_iqn": "", "ansible_kernel": "4.18.0-80.el8.x86_64", "ansible_lo": {"active": true, "device": "lo", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "on", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "on [fixed]", "netns_local": "on [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "on [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "off [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "on", "tcp_segmentation_offload": "on", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "tx_checksumming": "on", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "on [fixed]", "tx_nocache_copy": "off [fixed]", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tx_sctp_segmentation": "on", "tx_tcp6_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp_segmentation": "on", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "on [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "127.0.0.1", "broadcast": "host", "netmask": "255.0.0.0", "network": "127.0.0.0"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "mtu": 65536, "promisc": false, "timestamping": ["tx_software", "rx_software", "software"], "type": "loopback"}, "ansible_local": {}, "ansible_lsb": {}, "ansible_machine": "x86_64", "ansible_machine_id": "8ed79b23e8944b77a81448b8c18d6385", "ansible_memfree_mb": 386, "ansible_memory_mb": {"nocache": {"free": 624, "used": 196}, "real": {"free": 386, "total": 820, "used": 434}, "swap": {"cached": 0, "free": 0, "total": 0, "used": 0}}, "ansible_memtotal_mb": 820, "ansible_mounts": [{"block_available": 2290095, "block_size": 4096, "block_total": 2618619, "block_used": 328524, "device": "/dev/vda1", "fstype": "xfs", "inode_available": 5207720, "inode_total": 5242352, "inode_used": 34632, "mount": "/", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 9380229120, "size_total": 10725863424, "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}], "ansible_nodename": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_os_family": "RedHat", "ansible_pkg_mgr": "dnf", "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": ["ttyS0", "ttyS0,115200n8"], "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_processor": ["0", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "1", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "2", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz"], "ansible_processor_cores": 1, "ansible_processor_count": 3, "ansible_processor_threads_per_core": 1, "ansible_processor_vcpus": 3, "ansible_product_name": "Standard PC (i440FX + PIIX, 1996)", "ansible_product_serial": "NA", "ansible_product_uuid": "NA", "ansible_product_version": "pc-i440fx-3.1", "ansible_python": {"executable": "/usr/libexec/platform-python", "has_sslcontext": true, "type": "cpython", "version": {"major": 3, "micro": 8, "minor": 6, "releaselevel": "final", "serial": 0}, "version_info": [3, 6, 8, "final", 0]}, "ansible_python_version": "3.6.8", "ansible_real_group_id": 0, "ansible_real_user_id": 0, "ansible_selinux": {"config_mode": "enforcing", "mode": "enforcing", "policyvers": 31, "status": "enabled", "type": "targeted"}, "ansible_selinux_python_present": true, "ansible_service_mgr": "systemd", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPrHhzAOIG6+LL5JbkuUkoS6QN4j5cOGJspsUV19zYh9DR5Zk7lWRqz5UKtQFuNGSfVRwLSFTeCrhP75k3gOiVA=", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIK3/mAvCYBfVsqqvNPNz7qufug5u8KiBJSu4z8AS9bXL", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABAQD6Go6LZTg9y8kSqXP2ROON9Y+wvSfDqG2VqC7z2AcPygXjqIsiijDmlyinPE3wuoIR18gBvxLV6u0KaK89dC7qDh1HBwaQJc8vuZ2va2Hkd+ySqksZglInllOwJ1hodTamnVnEq2Ldd90qjkQylvVewAmIijJLJNwfzyoIsav8xeamgT4FUYpJtZPeMMep3Fu+rqgjno2vxaTQz7Ey0VdW5ad1+/v7IDOyNu856y6xTnpJU3A35f4lDLYRT31LnSMf27zWOkz+JzZ+ltvuDPjVPv2Bla+cdkuc36NyVf8skGEPj2LGSI2yEAX/0RSBwVQxyvhIiPVaH6aA3/Eqtnal", "ansible_swapfree_mb": 0, "ansible_swaptotal_mb": 0, "ansible_system": "Linux", "ansible_system_capabilities": ["cap_chown", "cap_dac_override", "cap_dac_read_search", "cap_fowner", "cap_fsetid", "cap_kill", "cap_setgid", "cap_setuid", "cap_setpcap", "cap_linux_immutable", "cap_net_bind_service", "cap_net_broadcast", "cap_net_admin", "cap_net_raw", "cap_ipc_lock", "cap_ipc_owner", "cap_sys_module", "cap_sys_rawio", "cap_sys_chroot", "cap_sys_ptrace", "cap_sys_pacct", "cap_sys_admin", "cap_sys_boot", "cap_sys_nice", "cap_sys_resource", "cap_sys_time", "cap_sys_tty_config", "cap_mknod", "cap_lease", "cap_audit_write", "cap_audit_control", "cap_setfcap", "cap_mac_override", "cap_mac_admin", "cap_syslog", "cap_wake_alarm", "cap_block_suspend", "cap_audit_read+ep"], "ansible_system_capabilities_enforced": "True", "ansible_system_vendor": "QEMU", "ansible_uptime_seconds": 98, "ansible_user_dir": "/root", "ansible_user_gecos": "root", "ansible_user_gid": 0, "ansible_user_id": "root", "ansible_user_shell": "/bin/bash", "ansible_user_uid": 0, "ansible_userspace_architecture": "x86_64", "ansible_userspace_bits": "64", "ansible_virtualization_role": "guest", "ansible_virtualization_type": "kvm", "gather_subset": ["all"], "module_setup": true}, "changed": false} META: ran handlers TASK [include_role : storage] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_lvm_one_disk_multiple_volumes.yml:9 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "msg": "", "rc": 0, "results": ["Installed: python3-blivet", "Installed: libblockdev-part-2.19-7.el8.x86_64", "Installed: nss-3.41.0-5.el8.x86_64", "Installed: libblockdev-swap-2.19-7.el8.x86_64", "Installed: userspace-rcu-0.10.1-2.el8.x86_64", "Installed: libblockdev-utils-2.19-7.el8.x86_64", "Installed: mdadm-4.1-4.el8.x86_64", "Installed: nss-softokn-3.41.0-5.el8.x86_64", "Installed: libbytesize-1.4-1.el8.x86_64", "Installed: nss-softokn-freebl-3.41.0-5.el8.x86_64", "Installed: nss-sysinit-3.41.0-5.el8.x86_64", "Installed: nss-util-3.41.0-5.el8.x86_64", "Installed: volume_key-libs-0.3.11-4.el8.x86_64", "Installed: ndctl-62-2.el8.x86_64", "Installed: ndctl-libs-62-2.el8.x86_64", "Installed: blivet-data-1:3.1.0-9.el8.noarch", "Installed: python3-pyparted-1:3.11.0-13.el8.x86_64", "Installed: python3-blivet-1:3.1.0-9.el8.noarch", "Installed: python3-blockdev-2.19-7.el8.x86_64", "Installed: daxctl-libs-62-2.el8.x86_64", "Installed: gdisk-1.0.3-6.el8.x86_64", "Installed: python3-bytesize-1.4-1.el8.x86_64", "Installed: device-mapper-event-8:1.02.155-6.el8.x86_64", "Installed: device-mapper-event-libs-8:1.02.155-6.el8.x86_64", "Installed: libaio-0.3.110-12.el8.x86_64", "Installed: device-mapper-multipath-0.7.8-7.el8.x86_64", "Installed: device-mapper-multipath-libs-0.7.8-7.el8.x86_64", "Installed: device-mapper-persistent-data-0.7.6-1.el8.x86_64", "Installed: lsof-4.91-2.el8.x86_64", "Installed: libblockdev-2.19-7.el8.x86_64", "Installed: lvm2-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-crypto-2.19-7.el8.x86_64", "Installed: libblockdev-dm-2.19-7.el8.x86_64", "Installed: lvm2-libs-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-fs-2.19-7.el8.x86_64", "Installed: libblockdev-kbd-2.19-7.el8.x86_64", "Installed: libblockdev-loop-2.19-7.el8.x86_64", "Installed: libblockdev-lvm-2.19-7.el8.x86_64", "Installed: libblockdev-mdraid-2.19-7.el8.x86_64", "Installed: libblockdev-mpath-2.19-7.el8.x86_64", "Installed: nspr-4.20.0-2.el8.x86_64", "Installed: libblockdev-nvdimm-2.19-7.el8.x86_64"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": [], "changed": false, "leaves": [], "mounts": [], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [], "changed": false, "failed": false, "leaves": [], "mounts": [], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_lvm_one_disk_multiple_volumes.yml:12 included: /tmp/tmpfvho35gc/tests/get_unused_disk.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Find unused disks in the system] ********************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "disks": ["vdb"]} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:8 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"unused_disks": ["vdb"]}, "changed": false} TASK [Exit playbook when there's no unused disks in the system] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:13 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} META: TASK [Print unused disks] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:19 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "unused_disks": [ "vdb" ] } TASK [Create three LVM logical volumes under one volume group] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/tests_lvm_one_disk_multiple_volumes.yml:17 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "msg": "Nothing to do", "rc": 0, "results": ["Installed: python3-blivet"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '3g', 'mount_point': '/opt/test2'}, {'name': 'test3', 'size': '3g', 'mount_point': '/opt/test3'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "3g"}, {"mount_point": "/opt/test3", "name": "test3", "size": "3g"}]}]}, "ansible_loop_var": "pool", "changed": false, "pool": {"disks": ["vdb"], "name": "foo", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "3g"}, {"mount_point": "/opt/test3", "name": "test3", "size": "3g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '3g', 'mount_point': '/opt/test2'}, {'name': 'test3', 'size': '3g', 'mount_point': '/opt/test3'}]}, {'name': 'test1', 'size': '3g', 'mount_point': '/opt/test1'}]) => {"ansible_facts": {"vol_defaults": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}], "vol_pools": ["foo"], "vols": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}]}, "ansible_loop_var": "item", "changed": false, "item": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "3g"}, {"mount_point": "/opt/test3", "name": "test3", "size": "3g"}]}, {"mount_point": "/opt/test1", "name": "test1", "size": "3g"}]} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '3g', 'mount_point': '/opt/test2'}, {'name': 'test3', 'size': '3g', 'mount_point': '/opt/test3'}]}, {'name': 'test2', 'size': '3g', 'mount_point': '/opt/test2'}]) => {"ansible_facts": {"vol_defaults": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}], "vol_pools": ["foo", "foo"], "vols": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "3g"}]}, "ansible_loop_var": "item", "changed": false, "item": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "3g"}, {"mount_point": "/opt/test3", "name": "test3", "size": "3g"}]}, {"mount_point": "/opt/test2", "name": "test2", "size": "3g"}]} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '3g', 'mount_point': '/opt/test2'}, {'name': 'test3', 'size': '3g', 'mount_point': '/opt/test3'}]}, {'name': 'test3', 'size': '3g', 'mount_point': '/opt/test3'}]) => {"ansible_facts": {"vol_defaults": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}], "vol_pools": ["foo", "foo", "foo"], "vols": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "3g"}, {"mount_point": "/opt/test3", "name": "test3", "size": "3g"}]}, "ansible_loop_var": "item", "changed": false, "item": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "3g"}, {"mount_point": "/opt/test3", "name": "test3", "size": "3g"}]}, {"mount_point": "/opt/test3", "name": "test3", "size": "3g"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test1'}, {'state': 'present', 'type': 'lvm', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid'}]) => {"ansible_facts": {"all_vols": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}]}, "ansible_loop_var": "item", "changed": false, "idx": 0, "item": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}]} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'name': 'test2', 'size': '3g', 'mount_point': '/opt/test2'}, {'state': 'present', 'type': 'lvm', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid'}]) => {"ansible_facts": {"all_vols": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}]}, "ansible_loop_var": "item", "changed": false, "idx": 1, "item": [{"mount_point": "/opt/test2", "name": "test2", "size": "3g"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}]} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'name': 'test3', 'size': '3g', 'mount_point': '/opt/test3'}, {'state': 'present', 'type': 'lvm', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid'}]) => {"ansible_facts": {"all_vols": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}]}, "ansible_loop_var": "item", "changed": false, "idx": 2, "item": [{"mount_point": "/opt/test3", "name": "test3", "size": "3g"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '3g', 'mount_point': '/opt/test2'}, {'name': 'test3', 'size': '3g', 'mount_point': '/opt/test3'}]}) => {"ansible_facts": {"vols_by_pool": {"foo": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}]}}, "ansible_loop_var": "item", "changed": false, "item": {"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "3g"}, {"mount_point": "/opt/test3", "name": "test3", "size": "3g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '3g', 'mount_point': '/opt/test1'}, {'name': 'test2', 'size': '3g', 'mount_point': '/opt/test2'}, {'name': 'test3', 'size': '3g', 'mount_point': '/opt/test3'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}]}]}, "ansible_loop_var": "pool", "changed": false, "idx": 0, "pool": {"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "3g"}, {"mount_point": "/opt/test2", "name": "test2", "size": "3g"}, {"mount_point": "/opt/test3", "name": "test3", "size": "3g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}]}]}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [ { "disks": [ "vdb" ], "name": "foo", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" } ] } ] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": ["create format lvmpv on /dev/vdb", "create device /dev/foo", "create device /dev/mapper/foo-test3", "create format xfs on /dev/mapper/foo-test3", "create device /dev/mapper/foo-test2", "create format xfs on /dev/mapper/foo-test2", "create device /dev/mapper/foo-test1", "create format xfs on /dev/mapper/foo-test1"], "changed": true, "leaves": ["/dev/sr0", "/dev/vda1", "/dev/vdc", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2", "/dev/mapper/foo-test3"], "mounts": [{"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted"}, {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted"}, {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "mounted"}], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [ "create format lvmpv on /dev/vdb", "create device /dev/foo", "create device /dev/mapper/foo-test3", "create format xfs on /dev/mapper/foo-test3", "create device /dev/mapper/foo-test2", "create format xfs on /dev/mapper/foo-test2", "create device /dev/mapper/foo-test1", "create format xfs on /dev/mapper/foo-test1" ], "changed": true, "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vdc", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2", "/dev/mapper/foo-test3" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted" }, { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "mounted" } ], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted"}, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1"} changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'src': '/dev/mapper/foo-test2', 'path': '/opt/test2', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/foo-test2", "state": "mounted"}, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test2"} changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'src': '/dev/mapper/foo-test3', 'path': '/opt/test3', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test3", "src": "/dev/mapper/foo-test3", "state": "mounted"}, "name": "/opt/test3", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test3"} TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "cmd": ["systemctl", "daemon-reload"], "delta": "0:00:00.720667", "end": "2019-05-26 06:48:37.165596", "rc": 0, "start": "2019-05-26 06:48:36.444929", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_lvm_one_disk_multiple_volumes.yml:35 included: /tmp/tmpfvho35gc/tests/verify-role-results.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:1 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"storage_tests_list": ["verify-volumes"]}, "changed": false} TASK [Print out pool information] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:4 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "_storage_pools_list": [ { "disks": [ "vdb" ], "name": "foo", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" } ] } ] } TASK [Print out volume information] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:9 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} TASK [Ensure the specified devices in the role were correctly configured in the user's system] ************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:14 included: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Read the /etc/fstab file for volume existence] ******************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "cmd": ["cat", "/etc/fstab"], "delta": "0:00:00.010094", "end": "2019-05-26 06:48:41.794910", "rc": 0, "start": "2019-05-26 06:48:41.784816", "stderr": "", "stderr_lines": [], "stdout": "\n#\n# /etc/fstab\n# Created by anaconda on Thu Apr 4 08:39:20 2019\n#\n# Accessible filesystems, by reference, are maintained under '/dev/disk/'.\n# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.\n#\n# After editing this file, run 'systemctl daemon-reload' to update systemd\n# units generated from this file.\n#\nUUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0\n/dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0\n/dev/mapper/foo-test2 /opt/test2 xfs defaults 0 0\n/dev/mapper/foo-test3 /opt/test3 xfs defaults 0 0", "stdout_lines": ["", "#", "# /etc/fstab", "# Created by anaconda on Thu Apr 4 08:39:20 2019", "#", "# Accessible filesystems, by reference, are maintained under '/dev/disk/'.", "# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.", "#", "# After editing this file, run 'systemctl daemon-reload' to update systemd", "# units generated from this file.", "#", "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0", "/dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0", "/dev/mapper/foo-test2 /opt/test2 xfs defaults 0 0", "/dev/mapper/foo-test3 /opt/test3 xfs defaults 0 0"]} TASK [Verify the volumes listed in storage_pools were correctly managed] *********************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:7 included: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"mount_entries": []}, "changed": false} TASK [gather mounts related to each volume in the pool] **************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => {"ansible_facts": {"mount_entries": [[{"block_available": 770128, "block_size": 4096, "block_total": 783872, "block_used": 13744, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3154444288, "size_total": 3210739712, "uuid": "ca761ec0-7228-4209-8c18-5fb815671854"}]]}, "ansible_loop_var": "current_volume", "changed": false, "current_volume": {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test2', 'pool': 'foo'}) => {"ansible_facts": {"mount_entries": [[{"block_available": 770128, "block_size": 4096, "block_total": 783872, "block_used": 13744, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3154444288, "size_total": 3210739712, "uuid": "ca761ec0-7228-4209-8c18-5fb815671854"}], [{"block_available": 770128, "block_size": 4096, "block_total": 783872, "block_used": 13744, "device": "/dev/mapper/foo-test2", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3154444288, "size_total": 3210739712, "uuid": "dc67d8a1-4507-4e5c-ad41-25640596e5fe"}]]}, "ansible_loop_var": "current_volume", "changed": false, "current_volume": {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test3', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test3', 'pool': 'foo'}) => {"ansible_facts": {"mount_entries": [[{"block_available": 770128, "block_size": 4096, "block_total": 783872, "block_used": 13744, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3154444288, "size_total": 3210739712, "uuid": "ca761ec0-7228-4209-8c18-5fb815671854"}], [{"block_available": 770128, "block_size": 4096, "block_total": 783872, "block_used": 13744, "device": "/dev/mapper/foo-test2", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3154444288, "size_total": 3210739712, "uuid": "dc67d8a1-4507-4e5c-ad41-25640596e5fe"}], [{"block_available": 770128, "block_size": 4096, "block_total": 783872, "block_used": 13744, "device": "/dev/mapper/foo-test3", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test3", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 3154444288, "size_total": 3210739712, "uuid": "e8e4cd32-2faa-4958-9505-eec6d4f8c974"}]]}, "ansible_loop_var": "current_volume", "changed": false, "current_volume": {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "pool": "foo", "size": "3g", "state": "present", "type": "lvm"}} TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:12 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test2', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, "idx": 1, "msg": "All assertions passed" } ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test3', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test3', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, "idx": 2, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:25 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test2', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, "idx": 1, "msg": "All assertions passed" } ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test3', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test3', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, "idx": 2, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab.] **************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:38 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, "msg": "All assertions passed" } ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test2', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test2', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "test2", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, "msg": "All assertions passed" } ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '3g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test3', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test3', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test3", "name": "test3", "pool": "foo", "size": "3g", "state": "present", "type": "lvm" }, "msg": "All assertions passed" } TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:14 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test1", "device": "/dev/mapper/foo-test1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test2) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test2", "device": "/dev/mapper/foo-test2"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test3) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test3", "device": "/dev/mapper/foo-test3"} TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:21 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vda1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vda1", "device": "/dev/vda1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test1", "device": "/dev/mapper/foo-test1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test2) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test2", "device": "/dev/mapper/foo-test2"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test3) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test3", "device": "/dev/mapper/foo-test3"} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:28 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_mount_devs": ["/dev/vda1", "/dev/mapper/foo-test1", "/dev/mapper/foo-test2", "/dev/mapper/foo-test3"], "blivet_mount_devs": ["/dev/mapper/foo-test1", "/dev/mapper/foo-test2", "/dev/mapper/foo-test3"], "removed_volumes": []}, "changed": false} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:33 TASK [Verify the volumes with no pool were correctly managed] ********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:40 TASK [Verify the present volumes with no pool were correctly managed] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:52 TASK [Verify the absent volumes with no pool are not in ansible_mounts] ************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:65 META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=52 changed=5 unreachable=0 failed=0 skipped=20 rescued=0 ignored=0 + cd /tmp/tmpfvho35gc/tests; TEST_SUBJECTS=/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TEST_ARTIFACTS=/tmp/linux-system-role-test-work-pull-linux-system-roles_storage-25-db52957-rhel-8-df0ousve/artifacts ansible-playbook -vv --inventory=/usr/share/ansible/inventory/standard-inventory-qcow2 /tmp/tmpfvho35gc/_setup.yml /tmp/tmpfvho35gc/tests/tests_lvm_one_disk_one_volume.yml ansible-playbook 2.8.0 config file = /home/till/.ansible.cfg configured module search path = ['/home/till/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 3.7.3 (default, May 11 2019, 00:38:04) [GCC 9.1.1 20190503 (Red Hat 9.1.1-1)] Using /home/till/.ansible.cfg as config file PLAYBOOK: _setup.yml *************************************************************************************************************************************************************************************** 2 plays in /tmp/tmpfvho35gc/_setup.yml PLAY [Fail when only localhost is available] *************************************************************************************************************************************************************** META: ran handlers TASK [debug] *********************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/_setup.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "groups": { "all": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "localhost": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "subjects": [ "/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2" ], "ungrouped": [] } } TASK [fail] ************************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:7 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} META: ran handlers META: ran handlers PLAY [Setup repos] ***************************************************************************************************************************************************************************************** META: ran handlers TASK [set up internal repositories] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/_setup.yml:16 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", "changed": true} META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=2 changed=1 unreachable=0 failed=0 skipped=1 rescued=0 ignored=0 PLAYBOOK: tests_lvm_one_disk_one_volume.yml **************************************************************************************************************************************************************** 1 plays in /tmp/tmpfvho35gc/tests/tests_lvm_one_disk_one_volume.yml PLAY [all] ************************************************************************************************************************************************************************************************* TASK [Gathering Facts] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/tests_lvm_one_disk_one_volume.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_all_ipv4_addresses": ["10.0.2.15"], "ansible_all_ipv6_addresses": ["fec0::5054:ff:fe12:3456", "fe80::5054:ff:fe12:3456"], "ansible_apparmor": {"status": "disabled"}, "ansible_architecture": "x86_64", "ansible_bios_date": "04/01/2014", "ansible_bios_version": "1.12.0-2.fc30", "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": "ttyS0,115200n8", "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_date_time": {"date": "2019-05-26", "day": "26", "epoch": "1558867835", "hour": "06", "iso8601": "2019-05-26T10:50:35Z", "iso8601_basic": "20190526T065035805151", "iso8601_basic_short": "20190526T065035", "iso8601_micro": "2019-05-26T10:50:35.805398Z", "minute": "50", "month": "05", "second": "35", "time": "06:50:35", "tz": "EDT", "tz_offset": "-0400", "weekday": "Sonntag", "weekday_number": "0", "weeknumber": "20", "year": "2019"}, "ansible_default_ipv4": {"address": "10.0.2.15", "alias": "eth0", "broadcast": "10.0.2.255", "gateway": "10.0.2.2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "netmask": "255.255.255.0", "network": "10.0.2.0", "type": "ether"}, "ansible_default_ipv6": {"address": "fec0::5054:ff:fe12:3456", "gateway": "fe80::2", "interface": "eth0", "macaddress": "52:54:00:12:34:56", "mtu": 1500, "prefix": "64", "scope": "site", "type": "ether"}, "ansible_device_links": {"ids": {"sr0": ["ata-QEMU_DVD-ROM_QM00003"]}, "labels": {"sr0": ["cidata"]}, "masters": {}, "uuids": {"sr0": ["2019-05-26-12-48-53-00"], "vda1": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}}, "ansible_devices": {"sr0": {"holders": [], "host": "IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]", "links": {"ids": ["ata-QEMU_DVD-ROM_QM00003"], "labels": ["cidata"], "masters": [], "uuids": ["2019-05-26-12-48-53-00"]}, "model": "QEMU DVD-ROM", "partitions": {}, "removable": "1", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "728", "sectorsize": "2048", "size": "364.00 KB", "support_discard": "0", "vendor": "QEMU", "virtual": 1}, "vda": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {"vda1": {"holders": [], "links": {"ids": [], "labels": [], "masters": [], "uuids": ["884f47c9-a69d-4c5b-915d-6b7c9c74c923"]}, "sectors": "20969439", "sectorsize": 512, "size": "10.00 GB", "start": "2048", "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdb": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}, "vdc": {"holders": [], "host": "SCSI storage controller: Red Hat, Inc. Virtio block device", "links": {"ids": [], "labels": [], "masters": [], "uuids": []}, "model": null, "partitions": {}, "removable": "0", "rotational": "1", "sas_address": null, "sas_device_handle": null, "scheduler_mode": "mq-deadline", "sectors": "20971520", "sectorsize": "512", "size": "10.00 GB", "support_discard": "0", "vendor": "0x1af4", "virtual": 1}}, "ansible_distribution": "RedHat", "ansible_distribution_file_parsed": true, "ansible_distribution_file_path": "/etc/redhat-release", "ansible_distribution_file_search_string": "Red Hat", "ansible_distribution_file_variety": "RedHat", "ansible_distribution_major_version": "8", "ansible_distribution_release": "Ootpa", "ansible_distribution_version": "8.0", "ansible_dns": {"nameservers": ["10.0.2.3"], "search": ["virt.pnr.lab.eng.rdu2.redhat.com"]}, "ansible_domain": "virt.pnr.lab.eng.rdu2.redhat.com", "ansible_effective_group_id": 0, "ansible_effective_user_id": 0, "ansible_env": {"DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "HOME": "/root", "LANG": "de_DE.UTF-8", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "LOGNAME": "root", "LS_COLORS": "rs=0:di=38;5;33:ln=38;5;51:mh=00:pi=40;38;5;11:so=38;5;13:do=38;5;5:bd=48;5;232;38;5;11:cd=48;5;232;38;5;3:or=48;5;232;38;5;9:mi=01;05;37;41:su=48;5;196;38;5;15:sg=48;5;11;38;5;16:ca=48;5;196;38;5;226:tw=48;5;10;38;5;16:ow=48;5;10;38;5;21:st=48;5;21;38;5;15:ex=38;5;40:*.tar=38;5;9:*.tgz=38;5;9:*.arc=38;5;9:*.arj=38;5;9:*.taz=38;5;9:*.lha=38;5;9:*.lz4=38;5;9:*.lzh=38;5;9:*.lzma=38;5;9:*.tlz=38;5;9:*.txz=38;5;9:*.tzo=38;5;9:*.t7z=38;5;9:*.zip=38;5;9:*.z=38;5;9:*.dz=38;5;9:*.gz=38;5;9:*.lrz=38;5;9:*.lz=38;5;9:*.lzo=38;5;9:*.xz=38;5;9:*.zst=38;5;9:*.tzst=38;5;9:*.bz2=38;5;9:*.bz=38;5;9:*.tbz=38;5;9:*.tbz2=38;5;9:*.tz=38;5;9:*.deb=38;5;9:*.rpm=38;5;9:*.jar=38;5;9:*.war=38;5;9:*.ear=38;5;9:*.sar=38;5;9:*.rar=38;5;9:*.alz=38;5;9:*.ace=38;5;9:*.zoo=38;5;9:*.cpio=38;5;9:*.7z=38;5;9:*.rz=38;5;9:*.cab=38;5;9:*.wim=38;5;9:*.swm=38;5;9:*.dwm=38;5;9:*.esd=38;5;9:*.jpg=38;5;13:*.jpeg=38;5;13:*.mjpg=38;5;13:*.mjpeg=38;5;13:*.gif=38;5;13:*.bmp=38;5;13:*.pbm=38;5;13:*.pgm=38;5;13:*.ppm=38;5;13:*.tga=38;5;13:*.xbm=38;5;13:*.xpm=38;5;13:*.tif=38;5;13:*.tiff=38;5;13:*.png=38;5;13:*.svg=38;5;13:*.svgz=38;5;13:*.mng=38;5;13:*.pcx=38;5;13:*.mov=38;5;13:*.mpg=38;5;13:*.mpeg=38;5;13:*.m2v=38;5;13:*.mkv=38;5;13:*.webm=38;5;13:*.ogm=38;5;13:*.mp4=38;5;13:*.m4v=38;5;13:*.mp4v=38;5;13:*.vob=38;5;13:*.qt=38;5;13:*.nuv=38;5;13:*.wmv=38;5;13:*.asf=38;5;13:*.rm=38;5;13:*.rmvb=38;5;13:*.flc=38;5;13:*.avi=38;5;13:*.fli=38;5;13:*.flv=38;5;13:*.gl=38;5;13:*.dl=38;5;13:*.xcf=38;5;13:*.xwd=38;5;13:*.yuv=38;5;13:*.cgm=38;5;13:*.emf=38;5;13:*.ogv=38;5;13:*.ogx=38;5;13:*.aac=38;5;45:*.au=38;5;45:*.flac=38;5;45:*.m4a=38;5;45:*.mid=38;5;45:*.midi=38;5;45:*.mka=38;5;45:*.mp3=38;5;45:*.mpc=38;5;45:*.ogg=38;5;45:*.ra=38;5;45:*.wav=38;5;45:*.oga=38;5;45:*.opus=38;5;45:*.spx=38;5;45:*.xspf=38;5;45:", "MAIL": "/var/mail/root", "PATH": "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "PWD": "/root", "SELINUX_LEVEL_REQUESTED": "", "SELINUX_ROLE_REQUESTED": "", "SELINUX_USE_CURRENT_RANGE": "", "SHELL": "/bin/bash", "SHLVL": "2", "SSH_CLIENT": "10.0.2.2 35654 22", "SSH_CONNECTION": "10.0.2.2 35654 10.0.2.15 22", "SSH_TTY": "/dev/pts/0", "TERM": "xterm-256color", "USER": "root", "XDG_RUNTIME_DIR": "/run/user/0", "XDG_SESSION_ID": "1", "_": "/usr/libexec/platform-python"}, "ansible_eth0": {"active": true, "device": "eth0", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "off [requested on]", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "off [fixed]", "netns_local": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "off [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "on [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "off", "tcp_segmentation_offload": "off", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "tx_checksumming": "off", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "off [fixed]", "tx_nocache_copy": "off", "tx_scatter_gather": "off [fixed]", "tx_scatter_gather_fraglist": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_tcp6_segmentation": "off [fixed]", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off [fixed]", "tx_tcp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "off [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "10.0.2.15", "broadcast": "10.0.2.255", "netmask": "255.255.255.0", "network": "10.0.2.0"}, "ipv6": [{"address": "fec0::5054:ff:fe12:3456", "prefix": "64", "scope": "site"}, {"address": "fe80::5054:ff:fe12:3456", "prefix": "64", "scope": "link"}], "macaddress": "52:54:00:12:34:56", "module": "virtio_net", "mtu": 1500, "pciid": "virtio0", "promisc": false, "speed": -1, "timestamping": ["tx_software", "rx_software", "software"], "type": "ether"}, "ansible_fibre_channel_wwn": [], "ansible_fips": false, "ansible_form_factor": "Other", "ansible_fqdn": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_hostname": "ibm-p8-kvm-03-guest-02", "ansible_hostnqn": "", "ansible_interfaces": ["eth0", "lo"], "ansible_is_chroot": false, "ansible_iscsi_iqn": "", "ansible_kernel": "4.18.0-80.el8.x86_64", "ansible_lo": {"active": true, "device": "lo", "features": {"esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "generic_receive_offload": "on", "generic_segmentation_offload": "on", "highdma": "on [fixed]", "hw_tc_offload": "off [fixed]", "l2_fwd_offload": "off [fixed]", "large_receive_offload": "off [fixed]", "loopback": "on [fixed]", "netns_local": "on [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "rx_all": "off [fixed]", "rx_checksumming": "on [fixed]", "rx_fcs": "off [fixed]", "rx_gro_hw": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "rx_vlan_filter": "off [fixed]", "rx_vlan_offload": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "scatter_gather": "on", "tcp_segmentation_offload": "on", "tls_hw_record": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "tx_checksumming": "on", "tx_esp_segmentation": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_gso_robust": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_lockless": "on [fixed]", "tx_nocache_copy": "off [fixed]", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tx_sctp_segmentation": "on", "tx_tcp6_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp_segmentation": "on", "tx_udp_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "udp_fragmentation_offload": "off", "vlan_challenged": "on [fixed]"}, "hw_timestamp_filters": [], "ipv4": {"address": "127.0.0.1", "broadcast": "host", "netmask": "255.0.0.0", "network": "127.0.0.0"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "mtu": 65536, "promisc": false, "timestamping": ["tx_software", "rx_software", "software"], "type": "loopback"}, "ansible_local": {}, "ansible_lsb": {}, "ansible_machine": "x86_64", "ansible_machine_id": "b988638c5c2a4835b4a491dff31ac44e", "ansible_memfree_mb": 387, "ansible_memory_mb": {"nocache": {"free": 625, "used": 195}, "real": {"free": 387, "total": 820, "used": 433}, "swap": {"cached": 0, "free": 0, "total": 0, "used": 0}}, "ansible_memtotal_mb": 820, "ansible_mounts": [{"block_available": 2290090, "block_size": 4096, "block_total": 2618619, "block_used": 328529, "device": "/dev/vda1", "fstype": "xfs", "inode_available": 5207720, "inode_total": 5242352, "inode_used": 34632, "mount": "/", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 9380208640, "size_total": 10725863424, "uuid": "884f47c9-a69d-4c5b-915d-6b7c9c74c923"}], "ansible_nodename": "ibm-p8-kvm-03-guest-02.virt.pnr.lab.eng.rdu2.redhat.com", "ansible_os_family": "RedHat", "ansible_pkg_mgr": "dnf", "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-80.el8.x86_64", "console": ["ttyS0", "ttyS0,115200n8"], "crashkernel": "auto", "net.ifnames": "0", "no_timer_check": true, "ro": true, "root": "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923"}, "ansible_processor": ["0", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "1", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz", "2", "GenuineIntel", "Intel(R) Core(TM) i7-7600U CPU @ 2.80GHz"], "ansible_processor_cores": 1, "ansible_processor_count": 3, "ansible_processor_threads_per_core": 1, "ansible_processor_vcpus": 3, "ansible_product_name": "Standard PC (i440FX + PIIX, 1996)", "ansible_product_serial": "NA", "ansible_product_uuid": "NA", "ansible_product_version": "pc-i440fx-3.1", "ansible_python": {"executable": "/usr/libexec/platform-python", "has_sslcontext": true, "type": "cpython", "version": {"major": 3, "micro": 8, "minor": 6, "releaselevel": "final", "serial": 0}, "version_info": [3, 6, 8, "final", 0]}, "ansible_python_version": "3.6.8", "ansible_real_group_id": 0, "ansible_real_user_id": 0, "ansible_selinux": {"config_mode": "enforcing", "mode": "enforcing", "policyvers": 31, "status": "enabled", "type": "targeted"}, "ansible_selinux_python_present": true, "ansible_service_mgr": "systemd", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLw28TKIgHjVZeZ8H4mT/h1VikTp57kIlvR4klXNZtsLEw2qcST7s2yNiu8HFBg9E/oB1L6iLLr3MWp7BeEOpFE=", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIAihuEXrRLmCdfROnql9ohqPEC1fYCX1PIvmyTqS1il8", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABAQDOrFVRM01ZJa4isPZ1qfWLZF1XvBKnoCvXGaugr7m3eYBB07kmTUi5La3OyEjUfIGz6QA2a9/U8sJZvblJKie6+p2brP2E5CcWC8mXdXNy90BKg1jPhgsJ6p4qg0oVKL8Xtmhnax5anbXlDHsrJI+tqviGRmW2T/mmolujve9SLsFtLW/3MfLTfxi/VaNe1B52bgVFY/pRR6k+QN5FwXo95oTZy9Xnn6SJcdREuPVhBqxfF0ru2OfT97CMln0e0GrX01kehLJG66WXmt4cbw0sw9sk2NbUTzyIUVaKoryr6iizM5Fjt4EHyEhjTk/wADZ+pF3WFGiqxr5R5Tp8e3Fn", "ansible_swapfree_mb": 0, "ansible_swaptotal_mb": 0, "ansible_system": "Linux", "ansible_system_capabilities": ["cap_chown", "cap_dac_override", "cap_dac_read_search", "cap_fowner", "cap_fsetid", "cap_kill", "cap_setgid", "cap_setuid", "cap_setpcap", "cap_linux_immutable", "cap_net_bind_service", "cap_net_broadcast", "cap_net_admin", "cap_net_raw", "cap_ipc_lock", "cap_ipc_owner", "cap_sys_module", "cap_sys_rawio", "cap_sys_chroot", "cap_sys_ptrace", "cap_sys_pacct", "cap_sys_admin", "cap_sys_boot", "cap_sys_nice", "cap_sys_resource", "cap_sys_time", "cap_sys_tty_config", "cap_mknod", "cap_lease", "cap_audit_write", "cap_audit_control", "cap_setfcap", "cap_mac_override", "cap_mac_admin", "cap_syslog", "cap_wake_alarm", "cap_block_suspend", "cap_audit_read+ep"], "ansible_system_capabilities_enforced": "True", "ansible_system_vendor": "QEMU", "ansible_uptime_seconds": 98, "ansible_user_dir": "/root", "ansible_user_gecos": "root", "ansible_user_gid": 0, "ansible_user_id": "root", "ansible_user_shell": "/bin/bash", "ansible_user_uid": 0, "ansible_userspace_architecture": "x86_64", "ansible_userspace_bits": "64", "ansible_virtualization_role": "guest", "ansible_virtualization_type": "kvm", "gather_subset": ["all"], "module_setup": true}, "changed": false} META: ran handlers TASK [include_role : storage] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_lvm_one_disk_one_volume.yml:10 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "msg": "", "rc": 0, "results": ["Installed: python3-blivet", "Installed: libblockdev-part-2.19-7.el8.x86_64", "Installed: nss-3.41.0-5.el8.x86_64", "Installed: libblockdev-swap-2.19-7.el8.x86_64", "Installed: userspace-rcu-0.10.1-2.el8.x86_64", "Installed: libblockdev-utils-2.19-7.el8.x86_64", "Installed: mdadm-4.1-4.el8.x86_64", "Installed: nss-softokn-3.41.0-5.el8.x86_64", "Installed: libbytesize-1.4-1.el8.x86_64", "Installed: nss-softokn-freebl-3.41.0-5.el8.x86_64", "Installed: nss-sysinit-3.41.0-5.el8.x86_64", "Installed: nss-util-3.41.0-5.el8.x86_64", "Installed: volume_key-libs-0.3.11-4.el8.x86_64", "Installed: ndctl-62-2.el8.x86_64", "Installed: ndctl-libs-62-2.el8.x86_64", "Installed: blivet-data-1:3.1.0-9.el8.noarch", "Installed: python3-pyparted-1:3.11.0-13.el8.x86_64", "Installed: python3-blivet-1:3.1.0-9.el8.noarch", "Installed: python3-blockdev-2.19-7.el8.x86_64", "Installed: daxctl-libs-62-2.el8.x86_64", "Installed: gdisk-1.0.3-6.el8.x86_64", "Installed: python3-bytesize-1.4-1.el8.x86_64", "Installed: device-mapper-event-8:1.02.155-6.el8.x86_64", "Installed: device-mapper-event-libs-8:1.02.155-6.el8.x86_64", "Installed: libaio-0.3.110-12.el8.x86_64", "Installed: device-mapper-multipath-0.7.8-7.el8.x86_64", "Installed: device-mapper-multipath-libs-0.7.8-7.el8.x86_64", "Installed: device-mapper-persistent-data-0.7.6-1.el8.x86_64", "Installed: lsof-4.91-2.el8.x86_64", "Installed: libblockdev-2.19-7.el8.x86_64", "Installed: lvm2-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-crypto-2.19-7.el8.x86_64", "Installed: libblockdev-dm-2.19-7.el8.x86_64", "Installed: lvm2-libs-8:2.03.02-6.el8.x86_64", "Installed: libblockdev-fs-2.19-7.el8.x86_64", "Installed: libblockdev-kbd-2.19-7.el8.x86_64", "Installed: libblockdev-loop-2.19-7.el8.x86_64", "Installed: libblockdev-lvm-2.19-7.el8.x86_64", "Installed: libblockdev-mdraid-2.19-7.el8.x86_64", "Installed: libblockdev-mpath-2.19-7.el8.x86_64", "Installed: nspr-4.20.0-2.el8.x86_64", "Installed: libblockdev-nvdimm-2.19-7.el8.x86_64"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": [], "changed": false, "leaves": [], "mounts": [], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [], "changed": false, "failed": false, "leaves": [], "mounts": [], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_lvm_one_disk_one_volume.yml:13 included: /tmp/tmpfvho35gc/tests/get_unused_disk.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Find unused disks in the system] ********************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "disks": ["vdb"]} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:8 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"unused_disks": ["vdb"]}, "changed": false} TASK [Exit playbook when there's no unused disks in the system] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:13 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} META: TASK [Print unused disks] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/get_unused_disk.yml:19 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "unused_disks": [ "vdb" ] } TASK [Create one LVM logical volume under one volume group] ************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/tests_lvm_one_disk_one_volume.yml:18 TASK [storage : define an empty list of pools to be used in testing] *************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": []}, "changed": false} TASK [storage : define an empty list of volumes to be used in testing] ************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main.yml:6 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : include the appropriate backend tasks] ***************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main.yml:10 included: /tmp/tmpfvho35gc/tasks/main-blivet.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [storage : get a list of rpm packages installed on host machine] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:2 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [storage : make sure blivet is available] ************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:7 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "msg": "Nothing to do", "rc": 0, "results": ["Installed: python3-blivet"]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:15 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"all_vols": [], "pools": [], "vol_pools": [], "vols": [], "vols_by_pool": [], "volumes": []}, "changed": false} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:23 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '4g', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "4g"}]}]}, "ansible_loop_var": "pool", "changed": false, "pool": {"disks": ["vdb"], "name": "foo", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "4g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:30 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '4g', 'mount_point': '/opt/test1'}]}, {'name': 'test1', 'size': '4g', 'mount_point': '/opt/test1'}]) => {"ansible_facts": {"vol_defaults": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}], "vol_pools": ["foo"], "vols": [{"mount_point": "/opt/test1", "name": "test1", "size": "4g"}]}, "ansible_loop_var": "item", "changed": false, "item": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "4g"}]}, {"mount_point": "/opt/test1", "name": "test1", "size": "4g"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:37 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=[{'name': 'test1', 'size': '4g', 'mount_point': '/opt/test1'}, {'state': 'present', 'type': 'lvm', 'size': 0, 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid'}]) => {"ansible_facts": {"all_vols": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "4g", "state": "present", "type": "lvm"}]}, "ansible_loop_var": "item", "changed": false, "idx": 0, "item": [{"mount_point": "/opt/test1", "name": "test1", "size": "4g"}, {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "", "size": 0, "state": "present", "type": "lvm"}]} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:44 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '4g', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"vols_by_pool": {"foo": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "4g", "state": "present", "type": "lvm"}]}}, "ansible_loop_var": "item", "changed": false, "item": {"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "4g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:49 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'name': 'foo', 'disks': ['vdb'], 'volumes': [{'name': 'test1', 'size': '4g', 'mount_point': '/opt/test1'}]}) => {"ansible_facts": {"pools": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "4g", "state": "present", "type": "lvm"}]}]}, "ansible_loop_var": "pool", "changed": false, "idx": 0, "pool": {"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"mount_point": "/opt/test1", "name": "test1", "size": "4g"}]}} TASK [storage : set_fact] ********************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:57 TASK [storage : set the list of pools for test verification] *********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:72 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_pools_list": [{"disks": ["vdb"], "name": "foo", "state": "present", "type": "lvm", "volumes": [{"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "4g", "state": "present", "type": "lvm"}]}]}, "changed": false} TASK [storage : set the list of volumes for test verification] ********************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:76 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"_storage_volumes_list": []}, "changed": false} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:80 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "pools": [ { "disks": [ "vdb" ], "name": "foo", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "4g", "state": "present", "type": "lvm" } ] } ] } TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:83 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "volumes": [] } TASK [storage : manage the pools] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:86 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"actions": ["create format lvmpv on /dev/vdb", "create device /dev/foo", "create device /dev/mapper/foo-test1", "create format xfs on /dev/mapper/foo-test1"], "changed": true, "leaves": ["/dev/sr0", "/dev/vda1", "/dev/vdc", "/dev/mapper/foo-test1"], "mounts": [{"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted"}], "removed_mounts": []} TASK [storage : debug] ************************************************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:92 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "blivet_output": { "actions": [ "create format lvmpv on /dev/vdb", "create device /dev/foo", "create device /dev/mapper/foo-test1", "create format xfs on /dev/mapper/foo-test1" ], "changed": true, "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vdc", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "removed_mounts": [] } } TASK [storage : remove obsolete mounts] ******************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:103 TASK [storage : add new mounts] **************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:111 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => {"ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": {"dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted"}, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1"} TASK [storage : tell systemd to refresh its view of /etc/fstab] ******************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:122 changed: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": true, "cmd": ["systemctl", "daemon-reload"], "delta": "0:00:00.680556", "end": "2019-05-26 06:51:47.552359", "rc": 0, "start": "2019-05-26 06:51:46.871803", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": []} TASK [storage : Update facts] ****************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tasks/main-blivet.yml:131 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] TASK [include_tasks] *************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/tests_lvm_one_disk_one_volume.yml:30 included: /tmp/tmpfvho35gc/tests/verify-role-results.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:1 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"storage_tests_list": ["verify-volumes"]}, "changed": false} TASK [Print out pool information] ************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:4 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => { "_storage_pools_list": [ { "disks": [ "vdb" ], "name": "foo", "state": "present", "type": "lvm", "volumes": [ { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "4g", "state": "present", "type": "lvm" } ] } ] } TASK [Print out volume information] ************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:9 skipping: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {} TASK [Ensure the specified devices in the role were correctly configured in the user's system] ************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/verify-role-results.yml:14 included: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [Read the /etc/fstab file for volume existence] ******************************************************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"changed": false, "cmd": ["cat", "/etc/fstab"], "delta": "0:00:00.003289", "end": "2019-05-26 06:51:52.010089", "rc": 0, "start": "2019-05-26 06:51:52.006800", "stderr": "", "stderr_lines": [], "stdout": "\n#\n# /etc/fstab\n# Created by anaconda on Thu Apr 4 08:39:20 2019\n#\n# Accessible filesystems, by reference, are maintained under '/dev/disk/'.\n# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.\n#\n# After editing this file, run 'systemctl daemon-reload' to update systemd\n# units generated from this file.\n#\nUUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0\n/dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0", "stdout_lines": ["", "#", "# /etc/fstab", "# Created by anaconda on Thu Apr 4 08:39:20 2019", "#", "# Accessible filesystems, by reference, are maintained under '/dev/disk/'.", "# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info.", "#", "# After editing this file, run 'systemctl daemon-reload' to update systemd", "# units generated from this file.", "#", "UUID=884f47c9-a69d-4c5b-915d-6b7c9c74c923 / xfs defaults 0 0", "/dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0"]} TASK [Verify the volumes listed in storage_pools were correctly managed] *********************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:7 included: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml for /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:2 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"mount_entries": []}, "changed": false} TASK [gather mounts related to each volume in the pool] **************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:5 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => {"ansible_facts": {"mount_entries": [[{"block_available": 1030436, "block_size": 4096, "block_total": 1046016, "block_used": 15580, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,noquota", "size_available": 4220665856, "size_total": 4284481536, "uuid": "215dc3f2-8730-4edc-ba23-85b00bbdc8c4"}]]}, "ansible_loop_var": "current_volume", "changed": false, "current_volume": {"fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "4g", "state": "present", "type": "lvm"}} TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:12 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "4g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab and ansible_mounts.] ********************************************************************************************************************* task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:25 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "4g", "state": "present", "type": "lvm" }, "idx": 0, "msg": "All assertions passed" } TASK [Ensure all lvm volumes were correctly managed in /etc/fstab.] **************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-volumes-present-helper.yml:38 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item={'state': 'present', 'type': 'lvm', 'size': '4g', 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_destroy_options': '-af', 'fs_overwrite_existing': True, 'mount_point': '/opt/test1', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'name': 'test1', 'pool': 'foo'}) => { "ansible_loop_var": "current_volume", "changed": false, "current_volume": { "fs_create_options": "", "fs_destroy_options": "-af", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "test1", "pool": "foo", "size": "4g", "state": "present", "type": "lvm" }, "msg": "All assertions passed" } TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:14 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test1", "device": "/dev/mapper/foo-test1"} TASK [resolve_blockdev] ************************************************************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:21 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/vda1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/vda1", "device": "/dev/vda1"} ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => (item=/dev/mapper/foo-test1) => {"ansible_loop_var": "dev_id", "changed": false, "dev_id": "/dev/mapper/foo-test1", "device": "/dev/mapper/foo-test1"} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:28 ok: [/home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2] => {"ansible_facts": {"ansible_mount_devs": ["/dev/vda1", "/dev/mapper/foo-test1"], "blivet_mount_devs": ["/dev/mapper/foo-test1"], "removed_volumes": []}, "changed": false} TASK [set_fact] ******************************************************************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:33 TASK [Verify the volumes with no pool were correctly managed] ********************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:40 TASK [Verify the present volumes with no pool were correctly managed] ************************************************************************************************************************************** task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:52 TASK [Verify the absent volumes with no pool are not in ansible_mounts] ************************************************************************************************************************************ task path: /tmp/tmpfvho35gc/tests/test-verify-volumes.yml:65 META: ran handlers META: ran handlers PLAY RECAP ************************************************************************************************************************************************************************************************* /home/till/lsr-test-harness-files/cache/rhel-guest-image-8.0-1854.x86_64.qcow2 : ok=52 changed=5 unreachable=0 failed=0 skipped=20 rescued=0 ignored=0