ansible-playbook 2.9.27 config file = /etc/ansible/ansible.cfg configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 2.7.5 (default, Nov 14 2023, 16:14:06) [GCC 4.8.5 20150623 (Red Hat 4.8.5-44)] Using /etc/ansible/ansible.cfg as config file Skipping callback 'actionable', as we already have a stdout callback. Skipping callback 'counter_enabled', as we already have a stdout callback. Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'full_skip', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'null', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. Skipping callback 'selective', as we already have a stdout callback. Skipping callback 'skippy', as we already have a stdout callback. Skipping callback 'stderr', as we already have a stdout callback. Skipping callback 'unixy', as we already have a stdout callback. Skipping callback 'yaml', as we already have a stdout callback. PLAYBOOK: tests_team_nm.yml **************************************************** 2 plays in /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/tests_team_nm.yml PLAY [Run playbook 'playbooks/tests_team.yml' with nm as provider] ************* TASK [Gathering Facts] ********************************************************* task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/tests_team_nm.yml:6 ok: [managed-node1] META: ran handlers TASK [Include the task 'el_repo_setup.yml'] ************************************ task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/tests_team_nm.yml:9 included: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml for managed-node1 TASK [Gather the minimum subset of ansible_facts required by the network role test] *** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:3 ok: [managed-node1] TASK [Check if system is ostree] *********************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:17 ok: [managed-node1] => {"changed": false, "stat": {"exists": false}} TASK [Set flag to indicate system is ostree] *********************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:22 ok: [managed-node1] => {"ansible_facts": {"__network_is_ostree": false}, "changed": false} TASK [Fix CentOS6 Base repo] *************************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:26 skipping: [managed-node1] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [Include the task 'enable_epel.yml'] ************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/tasks/el_repo_setup.yml:51 included: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml for managed-node1 TASK [Create EPEL 7] *********************************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml:8 ok: [managed-node1] => {"attempts": 1, "changed": false, "cmd": ["rpm", "-iv", "https://dl.fedoraproject.org/pub/archive/epel/7/x86_64/Packages/e/epel-release-7-14.noarch.rpm"], "rc": 0, "stdout": "skipped, since /etc/yum.repos.d/epel.repo exists", "stdout_lines": ["skipped, since /etc/yum.repos.d/epel.repo exists"]} TASK [Install yum-utils package] *********************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml:26 ok: [managed-node1] => {"changed": false, "msg": "", "rc": 0, "results": ["yum-utils-1.1.31-54.el7_8.noarch providing yum-utils is already installed"]} TASK [Enable EPEL 7] *********************************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml:32 ok: [managed-node1] => {"changed": false, "cmd": ["yum-config-manager", "--enable", "epel"], "delta": "0:00:00.167399", "end": "2024-09-24 14:49:56.383053", "rc": 0, "start": "2024-09-24 14:49:56.215654", "stderr": "", "stderr_lines": [], "stdout": "Loaded plugins: fastestmirror\n================================== repo: epel ==================================\n[epel]\nasync = True\nbandwidth = 0\nbase_persistdir = /var/lib/yum/repos/x86_64/7\nbaseurl = \ncache = 0\ncachedir = /var/cache/yum/x86_64/7/epel\ncheck_config_file_age = True\ncompare_providers_priority = 80\ncost = 1000\ndeltarpm_metadata_percentage = 100\ndeltarpm_percentage = \nenabled = True\nenablegroups = True\nexclude = \nfailovermethod = priority\nftp_disable_epsv = False\ngpgcadir = /var/lib/yum/repos/x86_64/7/epel/gpgcadir\ngpgcakey = \ngpgcheck = True\ngpgdir = /var/lib/yum/repos/x86_64/7/epel/gpgdir\ngpgkey = file:///etc/pki/rpm-gpg/RPM-GPG-KEY-EPEL-7\nhdrdir = /var/cache/yum/x86_64/7/epel/headers\nhttp_caching = all\nincludepkgs = \nip_resolve = \nkeepalive = True\nkeepcache = False\nmddownloadpolicy = sqlite\nmdpolicy = group:small\nmediaid = \nmetadata_expire = 21600\nmetadata_expire_filter = read-only:present\nmetalink = https://mirrors.fedoraproject.org/metalink?repo=epel-7&arch=x86_64&infra=stock&content=centos\nminrate = 0\nmirrorlist = \nmirrorlist_expire = 86400\nname = Extra Packages for Enterprise Linux 7 - x86_64\nold_base_cache_dir = \npassword = \npersistdir = /var/lib/yum/repos/x86_64/7/epel\npkgdir = /var/cache/yum/x86_64/7/epel/packages\nproxy = False\nproxy_dict = \nproxy_password = \nproxy_username = \nrepo_gpgcheck = False\nretries = 10\nskip_if_unavailable = False\nssl_check_cert_permissions = True\nsslcacert = \nsslclientcert = \nsslclientkey = \nsslverify = True\nthrottle = 0\ntimeout = 30.0\nui_id = epel/x86_64\nui_repoid_vars = releasever,\n basearch\nusername = ", "stdout_lines": ["Loaded plugins: fastestmirror", "================================== repo: epel ==================================", "[epel]", "async = True", "bandwidth = 0", "base_persistdir = /var/lib/yum/repos/x86_64/7", "baseurl = ", "cache = 0", "cachedir = /var/cache/yum/x86_64/7/epel", "check_config_file_age = True", "compare_providers_priority = 80", "cost = 1000", "deltarpm_metadata_percentage = 100", "deltarpm_percentage = ", "enabled = True", "enablegroups = True", "exclude = ", "failovermethod = priority", "ftp_disable_epsv = False", "gpgcadir = /var/lib/yum/repos/x86_64/7/epel/gpgcadir", "gpgcakey = ", "gpgcheck = True", "gpgdir = /var/lib/yum/repos/x86_64/7/epel/gpgdir", "gpgkey = file:///etc/pki/rpm-gpg/RPM-GPG-KEY-EPEL-7", "hdrdir = /var/cache/yum/x86_64/7/epel/headers", "http_caching = all", "includepkgs = ", "ip_resolve = ", "keepalive = True", "keepcache = False", "mddownloadpolicy = sqlite", "mdpolicy = group:small", "mediaid = ", "metadata_expire = 21600", "metadata_expire_filter = read-only:present", "metalink = https://mirrors.fedoraproject.org/metalink?repo=epel-7&arch=x86_64&infra=stock&content=centos", "minrate = 0", "mirrorlist = ", "mirrorlist_expire = 86400", "name = Extra Packages for Enterprise Linux 7 - x86_64", "old_base_cache_dir = ", "password = ", "persistdir = /var/lib/yum/repos/x86_64/7/epel", "pkgdir = /var/cache/yum/x86_64/7/epel/packages", "proxy = False", "proxy_dict = ", "proxy_password = ", "proxy_username = ", "repo_gpgcheck = False", "retries = 10", "skip_if_unavailable = False", "ssl_check_cert_permissions = True", "sslcacert = ", "sslclientcert = ", "sslclientkey = ", "sslverify = True", "throttle = 0", "timeout = 30.0", "ui_id = epel/x86_64", "ui_repoid_vars = releasever,", " basearch", "username = "]} TASK [Enable EPEL 8] *********************************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml:37 skipping: [managed-node1] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [Enable EPEL 6] *********************************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/tasks/enable_epel.yml:42 skipping: [managed-node1] => {"changed": false, "skip_reason": "Conditional result was False"} TASK [Set network provider to 'nm'] ******************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/tests_team_nm.yml:11 ok: [managed-node1] => {"ansible_facts": {"network_provider": "nm"}, "changed": false} META: ran handlers META: ran handlers PLAY [Play for testing team connection] **************************************** TASK [Gathering Facts] ********************************************************* task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_team.yml:3 ok: [managed-node1] META: ran handlers TASK [Show playbook name] ****************************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_team.yml:11 ok: [managed-node1] => { "msg": "this is: playbooks/tests_team.yml" } TASK [Include the task 'run_test.yml'] ***************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tests_team.yml:21 included: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml for managed-node1 TASK [TEST: Create a team interface without any port attached] ***************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:5 ok: [managed-node1] => { "msg": "########## Create a team interface without any port attached ##########" } TASK [Show item] *************************************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:9 ok: [managed-node1] => (item=lsr_description) => { "ansible_loop_var": "item", "item": "lsr_description", "lsr_description": "Create a team interface without any port attached" } ok: [managed-node1] => (item=lsr_setup) => { "ansible_loop_var": "item", "item": "lsr_setup", "lsr_setup": [ "tasks/delete_interface.yml", "tasks/assert_device_absent.yml" ] } ok: [managed-node1] => (item=lsr_test) => { "ansible_loop_var": "item", "item": "lsr_test", "lsr_test": [ "tasks/create_team_profile.yml" ] } ok: [managed-node1] => (item=lsr_assert) => { "ansible_loop_var": "item", "item": "lsr_assert", "lsr_assert": [ "tasks/assert_profile_present.yml", "tasks/assert_device_present.yml" ] } ok: [managed-node1] => (item=lsr_assert_when) => { "ansible_loop_var": "item", "item": "lsr_assert_when", "lsr_assert_when": "VARIABLE IS NOT DEFINED!: 'lsr_assert_when' is undefined" } ok: [managed-node1] => (item=lsr_fail_debug) => { "ansible_loop_var": "item", "item": "lsr_fail_debug", "lsr_fail_debug": [ "__network_connections_result" ] } ok: [managed-node1] => (item=lsr_cleanup) => { "ansible_loop_var": "item", "item": "lsr_cleanup", "lsr_cleanup": [ "tasks/cleanup_profile+device.yml", "tasks/check_network_dns.yml" ] } TASK [Include the task 'show_interfaces.yml'] ********************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:21 included: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/show_interfaces.yml for managed-node1 TASK [Include the task 'get_current_interfaces.yml'] *************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/show_interfaces.yml:3 included: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_current_interfaces.yml for managed-node1 TASK [Gather current interface info] ******************************************* task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_current_interfaces.yml:3 ok: [managed-node1] => {"changed": false, "cmd": ["ls", "-1"], "delta": "0:00:00.003230", "end": "2024-09-24 14:49:57.850126", "rc": 0, "start": "2024-09-24 14:49:57.846896", "stderr": "", "stderr_lines": [], "stdout": "eth0\nlo\npeertestnic1\ntest1\ntest1p\ntest2\ntest2p\ntestbr\ntestnic1", "stdout_lines": ["eth0", "lo", "peertestnic1", "test1", "test1p", "test2", "test2p", "testbr", "testnic1"]} TASK [Set current_interfaces] ************************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_current_interfaces.yml:9 ok: [managed-node1] => {"ansible_facts": {"current_interfaces": ["eth0", "lo", "peertestnic1", "test1", "test1p", "test2", "test2p", "testbr", "testnic1"]}, "changed": false} TASK [Show current_interfaces] ************************************************* task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/show_interfaces.yml:5 ok: [managed-node1] => { "msg": "current_interfaces: [u'eth0', u'lo', u'peertestnic1', u'test1', u'test1p', u'test2', u'test2p', u'testbr', u'testnic1']" } TASK [Setup] ******************************************************************* task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:24 included: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/delete_interface.yml for managed-node1 included: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_absent.yml for managed-node1 TASK [Remove test interface if necessary] ************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/delete_interface.yml:3 fatal: [managed-node1]: FAILED! => {"changed": false, "cmd": ["ip", "link", "del", "team0"], "delta": "0:00:00.006685", "end": "2024-09-24 14:49:58.344619", "msg": "non-zero return code", "rc": 1, "start": "2024-09-24 14:49:58.337934", "stderr": "Cannot find device \"team0\"", "stderr_lines": ["Cannot find device \"team0\""], "stdout": "", "stdout_lines": []} ...ignoring TASK [Include the task 'get_interface_stat.yml'] ******************************* task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_absent.yml:3 included: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml for managed-node1 TASK [Get stat for interface team0] ******************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/get_interface_stat.yml:3 ok: [managed-node1] => {"changed": false, "stat": {"exists": false}} TASK [Assert that the interface is absent - 'team0'] *************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/assert_device_absent.yml:5 ok: [managed-node1] => { "changed": false, "msg": "All assertions passed" } TASK [Test] ******************************************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:30 included: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_team_profile.yml for managed-node1 TASK [Include network role] **************************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_team_profile.yml:3 ERROR! the role 'fedora.linux_system_roles.network' was not found in /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/roles:/root/.ansible/roles:/usr/share/ansible/roles:/etc/ansible/roles:/tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks The error appears to be in '/tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/create_team_profile.yml': line 5, column 11, but may be elsewhere in the file depending on the exact syntax problem. The offending line appears to be: include_role: name: fedora.linux_system_roles.network ^ here TASK [Failure in test 'Create a team interface without any port attached'] ***** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:52 ok: [managed-node1] => { "msg": "!!!!! Failure in test 'Create a team interface without any port attached' !!!!!" } TASK [Show item that failed] *************************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:56 ok: [managed-node1] => (item=__network_connections_result) => { "__network_connections_result": "VARIABLE IS NOT DEFINED!: '__network_connections_result' is undefined", "ansible_loop_var": "item", "item": "__network_connections_result" } TASK [Issue failed message] **************************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:61 fatal: [managed-node1]: FAILED! => {"changed": false, "msg": "!!!!! Failure in test 'Create a team interface without any port attached' !!!!!"} TASK [Cleanup] ***************************************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/run_test.yml:66 included: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/cleanup_profile+device.yml for managed-node1 included: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/check_network_dns.yml for managed-node1 TASK [Cleanup profile and device] ********************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/cleanup_profile+device.yml:3 fatal: [managed-node1]: FAILED! => {"changed": false, "cmd": "nmcli con delete team0\nnmcli con load /etc/sysconfig/network-scripts/ifcfg-team0\nrm -f /etc/sysconfig/network-scripts/ifcfg-team0\nip link del team0\n", "delta": "0:00:00.050925", "end": "2024-09-24 14:49:59.915134", "msg": "non-zero return code", "rc": 1, "start": "2024-09-24 14:49:59.864209", "stderr": "Error: unknown connection 'team0'.\nError: cannot delete unknown connection(s): 'team0'.\nCould not load file '/etc/sysconfig/network-scripts/ifcfg-team0'\nCannot find device \"team0\"", "stderr_lines": ["Error: unknown connection 'team0'.", "Error: cannot delete unknown connection(s): 'team0'.", "Could not load file '/etc/sysconfig/network-scripts/ifcfg-team0'", "Cannot find device \"team0\""], "stdout": "", "stdout_lines": []} ...ignoring TASK [Check routes and DNS] **************************************************** task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/check_network_dns.yml:6 ok: [managed-node1] => {"changed": false, "cmd": "set -euo pipefail\necho IP\nip a\necho IP ROUTE\nip route\necho IP -6 ROUTE\nip -6 route\necho RESOLV\nif [ -f /etc/resolv.conf ]; then\n cat /etc/resolv.conf\nelse\n echo NO /etc/resolv.conf\n ls -alrtF /etc/resolv.* || :\nfi\n", "delta": "0:00:00.007007", "end": "2024-09-24 14:50:00.261873", "rc": 0, "start": "2024-09-24 14:50:00.254866", "stderr": "", "stderr_lines": [], "stdout": "IP\n1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000\n link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00\n inet 127.0.0.1/8 scope host lo\n valid_lft forever preferred_lft forever\n inet6 ::1/128 scope host \n valid_lft forever preferred_lft forever\n2: eth0: mtu 9001 qdisc mq state UP group default qlen 1000\n link/ether 0e:aa:93:f9:c8:c5 brd ff:ff:ff:ff:ff:ff\n inet 10.31.43.32/22 brd 10.31.43.255 scope global noprefixroute dynamic eth0\n valid_lft 3190sec preferred_lft 3190sec\n inet6 fe80::caa:93ff:fef9:c8c5/64 scope link \n valid_lft forever preferred_lft forever\n3: peertestnic1@testnic1: mtu 1500 qdisc noqueue state UP group default qlen 1000\n link/ether 26:53:8f:46:23:4a brd ff:ff:ff:ff:ff:ff\n inet6 fe80::2453:8fff:fe46:234a/64 scope link \n valid_lft forever preferred_lft forever\n4: testnic1@peertestnic1: mtu 1500 qdisc noqueue state UP group default qlen 1000\n link/ether 3a:fa:1e:4c:81:5e brd ff:ff:ff:ff:ff:ff\n inet6 fe80::5696:45ce:77c7:aa4c/64 scope link noprefixroute \n valid_lft forever preferred_lft forever\n5: test1p@test1: mtu 1500 qdisc noqueue master testbr state LOWERLAYERDOWN group default qlen 1000\n link/ether 76:bc:da:a1:9e:12 brd ff:ff:ff:ff:ff:ff\n6: test1@test1p: mtu 1500 qdisc noop state DOWN group default qlen 1000\n link/ether ca:25:c9:ae:60:b1 brd ff:ff:ff:ff:ff:ff\n7: test2p@test2: mtu 1500 qdisc noqueue master testbr state LOWERLAYERDOWN group default qlen 1000\n link/ether de:58:6b:39:c3:13 brd ff:ff:ff:ff:ff:ff\n8: test2@test2p: mtu 1500 qdisc noop state DOWN group default qlen 1000\n link/ether fa:98:ac:a0:b2:a2 brd ff:ff:ff:ff:ff:ff\n9: testbr: mtu 1500 qdisc noqueue state DOWN group default qlen 1000\n link/ether 76:bc:da:a1:9e:12 brd ff:ff:ff:ff:ff:ff\n inet 192.0.2.1/24 scope global testbr\n valid_lft forever preferred_lft forever\n inet6 2001:db8::74bc:daff:fea1:9e12/64 scope global mngtmpaddr dynamic \n valid_lft 1789sec preferred_lft 1789sec\n inet6 2001:db8::1/32 scope global \n valid_lft forever preferred_lft forever\n inet6 fe80::10d2:31ff:fe61:e3f9/64 scope link \n valid_lft forever preferred_lft forever\nIP ROUTE\ndefault via 10.31.40.1 dev eth0 proto dhcp metric 100 \n10.31.40.0/22 dev eth0 proto kernel scope link src 10.31.43.32 metric 100 \n192.0.2.0/24 dev testbr proto kernel scope link src 192.0.2.1 \nIP -6 ROUTE\nunreachable ::/96 dev lo metric 1024 error -113 pref medium\nunreachable ::ffff:0.0.0.0/96 dev lo metric 1024 error -113 pref medium\n2001:db8::/64 dev testbr proto kernel metric 256 expires 1788sec pref medium\n2001:db8::/32 dev testbr proto kernel metric 256 pref medium\nunreachable 2002:a00::/24 dev lo metric 1024 error -113 pref medium\nunreachable 2002:7f00::/24 dev lo metric 1024 error -113 pref medium\nunreachable 2002:a9fe::/32 dev lo metric 1024 error -113 pref medium\nunreachable 2002:ac10::/28 dev lo metric 1024 error -113 pref medium\nunreachable 2002:c0a8::/32 dev lo metric 1024 error -113 pref medium\nunreachable 2002:e000::/19 dev lo metric 1024 error -113 pref medium\nunreachable 3ffe:ffff::/32 dev lo metric 1024 error -113 pref medium\nfe80::/64 dev testnic1 proto kernel metric 101 pref medium\nfe80::/64 dev eth0 proto kernel metric 256 mtu 9001 pref medium\nfe80::/64 dev peertestnic1 proto kernel metric 256 pref medium\nfe80::/64 dev testbr proto kernel metric 256 pref medium\nRESOLV\n# Generated by NetworkManager\nsearch us-east-1.aws.redhat.com\nnameserver 10.29.169.13\nnameserver 10.29.170.12\nnameserver 10.2.32.1", "stdout_lines": ["IP", "1: lo: mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000", " link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00", " inet 127.0.0.1/8 scope host lo", " valid_lft forever preferred_lft forever", " inet6 ::1/128 scope host ", " valid_lft forever preferred_lft forever", "2: eth0: mtu 9001 qdisc mq state UP group default qlen 1000", " link/ether 0e:aa:93:f9:c8:c5 brd ff:ff:ff:ff:ff:ff", " inet 10.31.43.32/22 brd 10.31.43.255 scope global noprefixroute dynamic eth0", " valid_lft 3190sec preferred_lft 3190sec", " inet6 fe80::caa:93ff:fef9:c8c5/64 scope link ", " valid_lft forever preferred_lft forever", "3: peertestnic1@testnic1: mtu 1500 qdisc noqueue state UP group default qlen 1000", " link/ether 26:53:8f:46:23:4a brd ff:ff:ff:ff:ff:ff", " inet6 fe80::2453:8fff:fe46:234a/64 scope link ", " valid_lft forever preferred_lft forever", "4: testnic1@peertestnic1: mtu 1500 qdisc noqueue state UP group default qlen 1000", " link/ether 3a:fa:1e:4c:81:5e brd ff:ff:ff:ff:ff:ff", " inet6 fe80::5696:45ce:77c7:aa4c/64 scope link noprefixroute ", " valid_lft forever preferred_lft forever", "5: test1p@test1: mtu 1500 qdisc noqueue master testbr state LOWERLAYERDOWN group default qlen 1000", " link/ether 76:bc:da:a1:9e:12 brd ff:ff:ff:ff:ff:ff", "6: test1@test1p: mtu 1500 qdisc noop state DOWN group default qlen 1000", " link/ether ca:25:c9:ae:60:b1 brd ff:ff:ff:ff:ff:ff", "7: test2p@test2: mtu 1500 qdisc noqueue master testbr state LOWERLAYERDOWN group default qlen 1000", " link/ether de:58:6b:39:c3:13 brd ff:ff:ff:ff:ff:ff", "8: test2@test2p: mtu 1500 qdisc noop state DOWN group default qlen 1000", " link/ether fa:98:ac:a0:b2:a2 brd ff:ff:ff:ff:ff:ff", "9: testbr: mtu 1500 qdisc noqueue state DOWN group default qlen 1000", " link/ether 76:bc:da:a1:9e:12 brd ff:ff:ff:ff:ff:ff", " inet 192.0.2.1/24 scope global testbr", " valid_lft forever preferred_lft forever", " inet6 2001:db8::74bc:daff:fea1:9e12/64 scope global mngtmpaddr dynamic ", " valid_lft 1789sec preferred_lft 1789sec", " inet6 2001:db8::1/32 scope global ", " valid_lft forever preferred_lft forever", " inet6 fe80::10d2:31ff:fe61:e3f9/64 scope link ", " valid_lft forever preferred_lft forever", "IP ROUTE", "default via 10.31.40.1 dev eth0 proto dhcp metric 100 ", "10.31.40.0/22 dev eth0 proto kernel scope link src 10.31.43.32 metric 100 ", "192.0.2.0/24 dev testbr proto kernel scope link src 192.0.2.1 ", "IP -6 ROUTE", "unreachable ::/96 dev lo metric 1024 error -113 pref medium", "unreachable ::ffff:0.0.0.0/96 dev lo metric 1024 error -113 pref medium", "2001:db8::/64 dev testbr proto kernel metric 256 expires 1788sec pref medium", "2001:db8::/32 dev testbr proto kernel metric 256 pref medium", "unreachable 2002:a00::/24 dev lo metric 1024 error -113 pref medium", "unreachable 2002:7f00::/24 dev lo metric 1024 error -113 pref medium", "unreachable 2002:a9fe::/32 dev lo metric 1024 error -113 pref medium", "unreachable 2002:ac10::/28 dev lo metric 1024 error -113 pref medium", "unreachable 2002:c0a8::/32 dev lo metric 1024 error -113 pref medium", "unreachable 2002:e000::/19 dev lo metric 1024 error -113 pref medium", "unreachable 3ffe:ffff::/32 dev lo metric 1024 error -113 pref medium", "fe80::/64 dev testnic1 proto kernel metric 101 pref medium", "fe80::/64 dev eth0 proto kernel metric 256 mtu 9001 pref medium", "fe80::/64 dev peertestnic1 proto kernel metric 256 pref medium", "fe80::/64 dev testbr proto kernel metric 256 pref medium", "RESOLV", "# Generated by NetworkManager", "search us-east-1.aws.redhat.com", "nameserver 10.29.169.13", "nameserver 10.29.170.12", "nameserver 10.2.32.1"]} TASK [Verify DNS and network connectivity] ************************************* task path: /tmp/collections-aEW/ansible_collections/fedora/linux_system_roles/tests/network/playbooks/tasks/check_network_dns.yml:24 ok: [managed-node1] => {"changed": false, "cmd": "set -euo pipefail\necho CHECK DNS AND CONNECTIVITY\nfor host in mirrors.fedoraproject.org mirrors.centos.org; do\n if ! getent hosts \"$host\"; then\n echo FAILED to lookup host \"$host\"\n exit 1\n fi\n if ! curl -o /dev/null https://\"$host\"; then\n echo FAILED to contact host \"$host\"\n exit 1\n fi\ndone\n", "delta": "0:00:00.401803", "end": "2024-09-24 14:50:00.933119", "rc": 0, "start": "2024-09-24 14:50:00.531316", "stderr": " % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n\r 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0\r 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0\r100 305 100 305 0 0 2279 0 --:--:-- --:--:-- --:--:-- 2276\n % Total % Received % Xferd Average Speed Time Time Time Current\n Dload Upload Total Spent Left Speed\n\r 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0\r100 291 100 291 0 0 1171 0 --:--:-- --:--:-- --:--:-- 1173", "stderr_lines": [" % Total % Received % Xferd Average Speed Time Time Time Current", " Dload Upload Total Spent Left Speed", "", " 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0", " 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0", "100 305 100 305 0 0 2279 0 --:--:-- --:--:-- --:--:-- 2276", " % Total % Received % Xferd Average Speed Time Time Time Current", " Dload Upload Total Spent Left Speed", "", " 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0", "100 291 100 291 0 0 1171 0 --:--:-- --:--:-- --:--:-- 1173"], "stdout": "CHECK DNS AND CONNECTIVITY\n2620:52:3:1:dead:beef:cafe:fed6 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2600:2701:4000:5211:dead:beef:fe:fed3 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2605:bc80:3010:600:dead:beef:cafe:fed9 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2600:1f14:fad:5c02:7c8a:72d0:1c58:c189 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2604:1580:fe00:0:dead:beef:cafe:fed1 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed7 wildcard.fedoraproject.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed6 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2605:bc80:3010:600:dead:beef:cafe:fed9 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2600:2701:4000:5211:dead:beef:fe:fed3 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2604:1580:fe00:0:dead:beef:cafe:fed1 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2600:1f14:fad:5c02:7c8a:72d0:1c58:c189 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org\n2620:52:3:1:dead:beef:cafe:fed7 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org", "stdout_lines": ["CHECK DNS AND CONNECTIVITY", "2620:52:3:1:dead:beef:cafe:fed6 wildcard.fedoraproject.org mirrors.fedoraproject.org", "2600:2701:4000:5211:dead:beef:fe:fed3 wildcard.fedoraproject.org mirrors.fedoraproject.org", "2605:bc80:3010:600:dead:beef:cafe:fed9 wildcard.fedoraproject.org mirrors.fedoraproject.org", "2600:1f14:fad:5c02:7c8a:72d0:1c58:c189 wildcard.fedoraproject.org mirrors.fedoraproject.org", "2604:1580:fe00:0:dead:beef:cafe:fed1 wildcard.fedoraproject.org mirrors.fedoraproject.org", "2620:52:3:1:dead:beef:cafe:fed7 wildcard.fedoraproject.org mirrors.fedoraproject.org", "2620:52:3:1:dead:beef:cafe:fed6 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org", "2605:bc80:3010:600:dead:beef:cafe:fed9 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org", "2600:2701:4000:5211:dead:beef:fe:fed3 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org", "2604:1580:fe00:0:dead:beef:cafe:fed1 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org", "2600:1f14:fad:5c02:7c8a:72d0:1c58:c189 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org", "2620:52:3:1:dead:beef:cafe:fed7 wildcard.fedoraproject.org mirrors.centos.org mirrors.fedoraproject.org"]} PLAY RECAP ********************************************************************* managed-node1 : ok=34 changed=0 unreachable=0 failed=1 skipped=3 rescued=0 ignored=2